struct ftrace_ops *op, struct pt_regs *pt_regs)
{
struct ftrace_probe_ops *probe_ops;
+ struct trace_array *tr = op->private;
probe_ops = container_of(op, struct ftrace_probe_ops, ops);
* on the hash. rcu_read_lock is too dangerous here.
*/
preempt_disable_notrace();
- probe_ops->func(ip, parent_ip, probe_ops, NULL);
+ probe_ops->func(ip, parent_ip, tr, probe_ops, NULL);
preempt_enable_notrace();
}
ops->ops.func = function_trace_probe_call;
ftrace_ops_init(&ops->ops);
INIT_LIST_HEAD(&ops->list);
+ ops->ops.private = tr;
}
mutex_lock(&ops->ops.func_hash->regex_lock);
* to give the caller an opportunity to do so.
*/
if (ops->init) {
- ret = ops->init(ops, entry->ip, data);
+ ret = ops->init(ops, tr, entry->ip, data);
if (ret < 0)
goto out;
}
hlist_for_each_entry(entry, &hash->buckets[i], hlist) {
if (ftrace_lookup_ip(old_hash, entry->ip))
continue;
- ops->free(ops, entry->ip, NULL);
+ ops->free(ops, tr, entry->ip, NULL);
}
}
goto out_unlock;
struct ftrace_hash *hash = NULL;
struct hlist_node *tmp;
struct hlist_head hhd;
+ struct trace_array *tr;
char str[KSYM_SYMBOL_LEN];
int i, ret;
int size;
if (!(ops->ops.flags & FTRACE_OPS_FL_INITIALIZED))
return -EINVAL;
+ tr = ops->ops.private;
+
if (glob && (strcmp(glob, "*") == 0 || !strlen(glob)))
func_g.search = NULL;
else if (glob) {
hlist_for_each_entry_safe(entry, tmp, &hhd, hlist) {
hlist_del(&entry->hlist);
if (ops->free)
- ops->free(ops, entry->ip, NULL);
+ ops->free(ops, tr, entry->ip, NULL);
kfree(entry);
}
mutex_unlock(&ftrace_lock);
#if defined(CONFIG_TRACER_SNAPSHOT) && defined(CONFIG_DYNAMIC_FTRACE)
static void
ftrace_snapshot(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
tracing_snapshot();
}
static void
ftrace_count_snapshot(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
long *count = NULL;
}
static int
-ftrace_snapshot_init(struct ftrace_probe_ops *ops, unsigned long ip,
- void *data)
+ftrace_snapshot_init(struct ftrace_probe_ops *ops, struct trace_array *tr,
+ unsigned long ip, void *data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
}
static void
-ftrace_snapshot_free(struct ftrace_probe_ops *ops, unsigned long ip,
- void **_data)
+ftrace_snapshot_free(struct ftrace_probe_ops *ops, struct trace_array *tr,
+ unsigned long ip, void **data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
struct list_head list;
void (*func)(unsigned long ip,
unsigned long parent_ip,
+ struct trace_array *tr,
struct ftrace_probe_ops *ops,
void **data);
int (*init)(struct ftrace_probe_ops *ops,
+ struct trace_array *tr,
unsigned long ip, void *data);
void (*free)(struct ftrace_probe_ops *ops,
+ struct trace_array *tr,
unsigned long ip, void **data);
int (*print)(struct seq_file *m,
unsigned long ip,
static void
event_enable_probe(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **_data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **_data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
struct event_probe_data *data;
static void
event_enable_count_probe(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **_data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **_data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
struct event_probe_data *data;
static int
event_enable_print(struct seq_file *m, unsigned long ip,
- struct ftrace_probe_ops *ops, void *_data)
+ struct ftrace_probe_ops *ops, void *_data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
struct event_probe_data *data;
}
static int
-event_enable_init(struct ftrace_probe_ops *ops, unsigned long ip,
- void *_data)
+event_enable_init(struct ftrace_probe_ops *ops, struct trace_array *tr,
+ unsigned long ip, void *_data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
struct event_probe_data *data = _data;
}
static void
-event_enable_free(struct ftrace_probe_ops *ops, unsigned long ip,
- void **_data)
+event_enable_free(struct ftrace_probe_ops *ops, struct trace_array *tr,
+ unsigned long ip, void **_data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
struct event_probe_data *data;
static void
ftrace_traceon_count(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
update_traceon_count(ops, ip, 1);
}
static void
ftrace_traceoff_count(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
update_traceon_count(ops, ip, 0);
}
static void
ftrace_traceon(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
if (tracing_is_on())
return;
static void
ftrace_traceoff(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
if (!tracing_is_on())
return;
static void
ftrace_stacktrace(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
trace_dump_stack(STACK_SKIP);
}
static void
ftrace_stacktrace_count(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
long *count;
static void
ftrace_dump_probe(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
if (update_count(ops, ip))
ftrace_dump(DUMP_ALL);
/* Only dump the current CPU buffer. */
static void
ftrace_cpudump_probe(unsigned long ip, unsigned long parent_ip,
- struct ftrace_probe_ops *ops, void **data)
+ struct trace_array *tr, struct ftrace_probe_ops *ops,
+ void **data)
{
if (update_count(ops, ip))
ftrace_dump(DUMP_ORIG);
static int
ftrace_traceon_print(struct seq_file *m, unsigned long ip,
- struct ftrace_probe_ops *ops, void *data)
+ struct ftrace_probe_ops *ops,
+ void *data)
{
return ftrace_probe_print("traceon", m, ip, ops);
}
static int
-ftrace_count_init(struct ftrace_probe_ops *ops, unsigned long ip,
- void *data)
+ftrace_count_init(struct ftrace_probe_ops *ops, struct trace_array *tr,
+ unsigned long ip, void *data)
{
struct ftrace_func_mapper *mapper = ops->private_data;
}
static void
-ftrace_count_free(struct ftrace_probe_ops *ops, unsigned long ip,
- void **_data)
+ftrace_count_free(struct ftrace_probe_ops *ops, struct trace_array *tr,
+ unsigned long ip, void **_data)
{
struct ftrace_func_mapper *mapper = ops->private_data;