1 // SPDX-License-Identifier: GPL-2.0
3 * ring buffer based function tracer
5 * Copyright (C) 2007-2012 Steven Rostedt <srostedt@redhat.com>
6 * Copyright (C) 2008 Ingo Molnar <mingo@redhat.com>
8 * Originally taken from the RT patch by:
9 * Arnaldo Carvalho de Melo <acme@redhat.com>
11 * Based on code from the latency_tracer, that is:
12 * Copyright (C) 2004-2006 Ingo Molnar
13 * Copyright (C) 2004 Nadia Yvette Chambers
15 #include <linux/ring_buffer.h>
16 #include <generated/utsrelease.h>
17 #include <linux/stacktrace.h>
18 #include <linux/writeback.h>
19 #include <linux/kallsyms.h>
20 #include <linux/security.h>
21 #include <linux/seq_file.h>
22 #include <linux/notifier.h>
23 #include <linux/irqflags.h>
24 #include <linux/debugfs.h>
25 #include <linux/tracefs.h>
26 #include <linux/pagemap.h>
27 #include <linux/hardirq.h>
28 #include <linux/linkage.h>
29 #include <linux/uaccess.h>
30 #include <linux/vmalloc.h>
31 #include <linux/ftrace.h>
32 #include <linux/module.h>
33 #include <linux/percpu.h>
34 #include <linux/splice.h>
35 #include <linux/kdebug.h>
36 #include <linux/string.h>
37 #include <linux/mount.h>
38 #include <linux/rwsem.h>
39 #include <linux/slab.h>
40 #include <linux/ctype.h>
41 #include <linux/init.h>
42 #include <linux/poll.h>
43 #include <linux/nmi.h>
45 #include <linux/trace.h>
46 #include <linux/sched/clock.h>
47 #include <linux/sched/rt.h>
48 #include <linux/fsnotify.h>
49 #include <linux/irq_work.h>
50 #include <linux/workqueue.h>
53 #include "trace_output.h"
56 * On boot up, the ring buffer is set to the minimum size, so that
57 * we do not waste memory on systems that are not using tracing.
59 bool ring_buffer_expanded
;
62 * We need to change this state when a selftest is running.
63 * A selftest will lurk into the ring-buffer to count the
64 * entries inserted during the selftest although some concurrent
65 * insertions into the ring-buffer such as trace_printk could occurred
66 * at the same time, giving false positive or negative results.
68 static bool __read_mostly tracing_selftest_running
;
71 * If a tracer is running, we do not want to run SELFTEST.
73 bool __read_mostly tracing_selftest_disabled
;
75 /* Pipe tracepoints to printk */
76 struct trace_iterator
*tracepoint_print_iter
;
77 int tracepoint_printk
;
78 static DEFINE_STATIC_KEY_FALSE(tracepoint_printk_key
);
80 /* For tracers that don't implement custom flags */
81 static struct tracer_opt dummy_tracer_opt
[] = {
86 dummy_set_flag(struct trace_array
*tr
, u32 old_flags
, u32 bit
, int set
)
92 * To prevent the comm cache from being overwritten when no
93 * tracing is active, only save the comm when a trace event
96 static DEFINE_PER_CPU(bool, trace_taskinfo_save
);
99 * Kill all tracing for good (never come back).
100 * It is initialized to 1 but will turn to zero if the initialization
101 * of the tracer is successful. But that is the only place that sets
104 static int tracing_disabled
= 1;
106 cpumask_var_t __read_mostly tracing_buffer_mask
;
109 * ftrace_dump_on_oops - variable to dump ftrace buffer on oops
111 * If there is an oops (or kernel panic) and the ftrace_dump_on_oops
112 * is set, then ftrace_dump is called. This will output the contents
113 * of the ftrace buffers to the console. This is very useful for
114 * capturing traces that lead to crashes and outputing it to a
117 * It is default off, but you can enable it with either specifying
118 * "ftrace_dump_on_oops" in the kernel command line, or setting
119 * /proc/sys/kernel/ftrace_dump_on_oops
120 * Set 1 if you want to dump buffers of all CPUs
121 * Set 2 if you want to dump the buffer of the CPU that triggered oops
124 enum ftrace_dump_mode ftrace_dump_on_oops
;
126 /* When set, tracing will stop when a WARN*() is hit */
127 int __disable_trace_on_warning
;
129 #ifdef CONFIG_TRACE_EVAL_MAP_FILE
130 /* Map of enums to their values, for "eval_map" file */
131 struct trace_eval_map_head
{
133 unsigned long length
;
136 union trace_eval_map_item
;
138 struct trace_eval_map_tail
{
140 * "end" is first and points to NULL as it must be different
141 * than "mod" or "eval_string"
143 union trace_eval_map_item
*next
;
144 const char *end
; /* points to NULL */
147 static DEFINE_MUTEX(trace_eval_mutex
);
150 * The trace_eval_maps are saved in an array with two extra elements,
151 * one at the beginning, and one at the end. The beginning item contains
152 * the count of the saved maps (head.length), and the module they
153 * belong to if not built in (head.mod). The ending item contains a
154 * pointer to the next array of saved eval_map items.
156 union trace_eval_map_item
{
157 struct trace_eval_map map
;
158 struct trace_eval_map_head head
;
159 struct trace_eval_map_tail tail
;
162 static union trace_eval_map_item
*trace_eval_maps
;
163 #endif /* CONFIG_TRACE_EVAL_MAP_FILE */
165 int tracing_set_tracer(struct trace_array
*tr
, const char *buf
);
166 static void ftrace_trace_userstack(struct trace_buffer
*buffer
,
167 unsigned long flags
, int pc
);
169 #define MAX_TRACER_SIZE 100
170 static char bootup_tracer_buf
[MAX_TRACER_SIZE
] __initdata
;
171 static char *default_bootup_tracer
;
173 static bool allocate_snapshot
;
175 static int __init
set_cmdline_ftrace(char *str
)
177 strlcpy(bootup_tracer_buf
, str
, MAX_TRACER_SIZE
);
178 default_bootup_tracer
= bootup_tracer_buf
;
179 /* We are using ftrace early, expand it */
180 ring_buffer_expanded
= true;
183 __setup("ftrace=", set_cmdline_ftrace
);
185 static int __init
set_ftrace_dump_on_oops(char *str
)
187 if (*str
++ != '=' || !*str
) {
188 ftrace_dump_on_oops
= DUMP_ALL
;
192 if (!strcmp("orig_cpu", str
)) {
193 ftrace_dump_on_oops
= DUMP_ORIG
;
199 __setup("ftrace_dump_on_oops", set_ftrace_dump_on_oops
);
201 static int __init
stop_trace_on_warning(char *str
)
203 if ((strcmp(str
, "=0") != 0 && strcmp(str
, "=off") != 0))
204 __disable_trace_on_warning
= 1;
207 __setup("traceoff_on_warning", stop_trace_on_warning
);
209 static int __init
boot_alloc_snapshot(char *str
)
211 allocate_snapshot
= true;
212 /* We also need the main ring buffer expanded */
213 ring_buffer_expanded
= true;
216 __setup("alloc_snapshot", boot_alloc_snapshot
);
219 static char trace_boot_options_buf
[MAX_TRACER_SIZE
] __initdata
;
221 static int __init
set_trace_boot_options(char *str
)
223 strlcpy(trace_boot_options_buf
, str
, MAX_TRACER_SIZE
);
226 __setup("trace_options=", set_trace_boot_options
);
228 static char trace_boot_clock_buf
[MAX_TRACER_SIZE
] __initdata
;
229 static char *trace_boot_clock __initdata
;
231 static int __init
set_trace_boot_clock(char *str
)
233 strlcpy(trace_boot_clock_buf
, str
, MAX_TRACER_SIZE
);
234 trace_boot_clock
= trace_boot_clock_buf
;
237 __setup("trace_clock=", set_trace_boot_clock
);
239 static int __init
set_tracepoint_printk(char *str
)
241 if ((strcmp(str
, "=0") != 0 && strcmp(str
, "=off") != 0))
242 tracepoint_printk
= 1;
245 __setup("tp_printk", set_tracepoint_printk
);
247 unsigned long long ns2usecs(u64 nsec
)
254 /* trace_flags holds trace_options default values */
255 #define TRACE_DEFAULT_FLAGS \
256 (FUNCTION_DEFAULT_FLAGS | \
257 TRACE_ITER_PRINT_PARENT | TRACE_ITER_PRINTK | \
258 TRACE_ITER_ANNOTATE | TRACE_ITER_CONTEXT_INFO | \
259 TRACE_ITER_RECORD_CMD | TRACE_ITER_OVERWRITE | \
260 TRACE_ITER_IRQ_INFO | TRACE_ITER_MARKERS)
262 /* trace_options that are only supported by global_trace */
263 #define TOP_LEVEL_TRACE_FLAGS (TRACE_ITER_PRINTK | \
264 TRACE_ITER_PRINTK_MSGONLY | TRACE_ITER_RECORD_CMD)
266 /* trace_flags that are default zero for instances */
267 #define ZEROED_TRACE_FLAGS \
268 (TRACE_ITER_EVENT_FORK | TRACE_ITER_FUNC_FORK)
271 * The global_trace is the descriptor that holds the top-level tracing
272 * buffers for the live tracing.
274 static struct trace_array global_trace
= {
275 .trace_flags
= TRACE_DEFAULT_FLAGS
,
278 LIST_HEAD(ftrace_trace_arrays
);
280 int trace_array_get(struct trace_array
*this_tr
)
282 struct trace_array
*tr
;
285 mutex_lock(&trace_types_lock
);
286 list_for_each_entry(tr
, &ftrace_trace_arrays
, list
) {
293 mutex_unlock(&trace_types_lock
);
298 static void __trace_array_put(struct trace_array
*this_tr
)
300 WARN_ON(!this_tr
->ref
);
305 * trace_array_put - Decrement the reference counter for this trace array.
307 * NOTE: Use this when we no longer need the trace array returned by
308 * trace_array_get_by_name(). This ensures the trace array can be later
312 void trace_array_put(struct trace_array
*this_tr
)
317 mutex_lock(&trace_types_lock
);
318 __trace_array_put(this_tr
);
319 mutex_unlock(&trace_types_lock
);
321 EXPORT_SYMBOL_GPL(trace_array_put
);
323 int tracing_check_open_get_tr(struct trace_array
*tr
)
327 ret
= security_locked_down(LOCKDOWN_TRACEFS
);
331 if (tracing_disabled
)
334 if (tr
&& trace_array_get(tr
) < 0)
340 int call_filter_check_discard(struct trace_event_call
*call
, void *rec
,
341 struct trace_buffer
*buffer
,
342 struct ring_buffer_event
*event
)
344 if (unlikely(call
->flags
& TRACE_EVENT_FL_FILTERED
) &&
345 !filter_match_preds(call
->filter
, rec
)) {
346 __trace_event_discard_commit(buffer
, event
);
353 void trace_free_pid_list(struct trace_pid_list
*pid_list
)
355 vfree(pid_list
->pids
);
360 * trace_find_filtered_pid - check if a pid exists in a filtered_pid list
361 * @filtered_pids: The list of pids to check
362 * @search_pid: The PID to find in @filtered_pids
364 * Returns true if @search_pid is fonud in @filtered_pids, and false otherwis.
367 trace_find_filtered_pid(struct trace_pid_list
*filtered_pids
, pid_t search_pid
)
370 * If pid_max changed after filtered_pids was created, we
371 * by default ignore all pids greater than the previous pid_max.
373 if (search_pid
>= filtered_pids
->pid_max
)
376 return test_bit(search_pid
, filtered_pids
->pids
);
380 * trace_ignore_this_task - should a task be ignored for tracing
381 * @filtered_pids: The list of pids to check
382 * @task: The task that should be ignored if not filtered
384 * Checks if @task should be traced or not from @filtered_pids.
385 * Returns true if @task should *NOT* be traced.
386 * Returns false if @task should be traced.
389 trace_ignore_this_task(struct trace_pid_list
*filtered_pids
,
390 struct trace_pid_list
*filtered_no_pids
,
391 struct task_struct
*task
)
394 * If filterd_no_pids is not empty, and the task's pid is listed
395 * in filtered_no_pids, then return true.
396 * Otherwise, if filtered_pids is empty, that means we can
397 * trace all tasks. If it has content, then only trace pids
398 * within filtered_pids.
401 return (filtered_pids
&&
402 !trace_find_filtered_pid(filtered_pids
, task
->pid
)) ||
404 trace_find_filtered_pid(filtered_no_pids
, task
->pid
));
408 * trace_filter_add_remove_task - Add or remove a task from a pid_list
409 * @pid_list: The list to modify
410 * @self: The current task for fork or NULL for exit
411 * @task: The task to add or remove
413 * If adding a task, if @self is defined, the task is only added if @self
414 * is also included in @pid_list. This happens on fork and tasks should
415 * only be added when the parent is listed. If @self is NULL, then the
416 * @task pid will be removed from the list, which would happen on exit
419 void trace_filter_add_remove_task(struct trace_pid_list
*pid_list
,
420 struct task_struct
*self
,
421 struct task_struct
*task
)
426 /* For forks, we only add if the forking task is listed */
428 if (!trace_find_filtered_pid(pid_list
, self
->pid
))
432 /* Sorry, but we don't support pid_max changing after setting */
433 if (task
->pid
>= pid_list
->pid_max
)
436 /* "self" is set for forks, and NULL for exits */
438 set_bit(task
->pid
, pid_list
->pids
);
440 clear_bit(task
->pid
, pid_list
->pids
);
444 * trace_pid_next - Used for seq_file to get to the next pid of a pid_list
445 * @pid_list: The pid list to show
446 * @v: The last pid that was shown (+1 the actual pid to let zero be displayed)
447 * @pos: The position of the file
449 * This is used by the seq_file "next" operation to iterate the pids
450 * listed in a trace_pid_list structure.
452 * Returns the pid+1 as we want to display pid of zero, but NULL would
453 * stop the iteration.
455 void *trace_pid_next(struct trace_pid_list
*pid_list
, void *v
, loff_t
*pos
)
457 unsigned long pid
= (unsigned long)v
;
461 /* pid already is +1 of the actual prevous bit */
462 pid
= find_next_bit(pid_list
->pids
, pid_list
->pid_max
, pid
);
464 /* Return pid + 1 to allow zero to be represented */
465 if (pid
< pid_list
->pid_max
)
466 return (void *)(pid
+ 1);
472 * trace_pid_start - Used for seq_file to start reading pid lists
473 * @pid_list: The pid list to show
474 * @pos: The position of the file
476 * This is used by seq_file "start" operation to start the iteration
479 * Returns the pid+1 as we want to display pid of zero, but NULL would
480 * stop the iteration.
482 void *trace_pid_start(struct trace_pid_list
*pid_list
, loff_t
*pos
)
487 pid
= find_first_bit(pid_list
->pids
, pid_list
->pid_max
);
488 if (pid
>= pid_list
->pid_max
)
491 /* Return pid + 1 so that zero can be the exit value */
492 for (pid
++; pid
&& l
< *pos
;
493 pid
= (unsigned long)trace_pid_next(pid_list
, (void *)pid
, &l
))
499 * trace_pid_show - show the current pid in seq_file processing
500 * @m: The seq_file structure to write into
501 * @v: A void pointer of the pid (+1) value to display
503 * Can be directly used by seq_file operations to display the current
506 int trace_pid_show(struct seq_file
*m
, void *v
)
508 unsigned long pid
= (unsigned long)v
- 1;
510 seq_printf(m
, "%lu\n", pid
);
514 /* 128 should be much more than enough */
515 #define PID_BUF_SIZE 127
517 int trace_pid_write(struct trace_pid_list
*filtered_pids
,
518 struct trace_pid_list
**new_pid_list
,
519 const char __user
*ubuf
, size_t cnt
)
521 struct trace_pid_list
*pid_list
;
522 struct trace_parser parser
;
530 if (trace_parser_get_init(&parser
, PID_BUF_SIZE
+ 1))
534 * Always recreate a new array. The write is an all or nothing
535 * operation. Always create a new array when adding new pids by
536 * the user. If the operation fails, then the current list is
539 pid_list
= kmalloc(sizeof(*pid_list
), GFP_KERNEL
);
541 trace_parser_put(&parser
);
545 pid_list
->pid_max
= READ_ONCE(pid_max
);
547 /* Only truncating will shrink pid_max */
548 if (filtered_pids
&& filtered_pids
->pid_max
> pid_list
->pid_max
)
549 pid_list
->pid_max
= filtered_pids
->pid_max
;
551 pid_list
->pids
= vzalloc((pid_list
->pid_max
+ 7) >> 3);
552 if (!pid_list
->pids
) {
553 trace_parser_put(&parser
);
559 /* copy the current bits to the new max */
560 for_each_set_bit(pid
, filtered_pids
->pids
,
561 filtered_pids
->pid_max
) {
562 set_bit(pid
, pid_list
->pids
);
571 ret
= trace_get_user(&parser
, ubuf
, cnt
, &pos
);
572 if (ret
< 0 || !trace_parser_loaded(&parser
))
580 if (kstrtoul(parser
.buffer
, 0, &val
))
582 if (val
>= pid_list
->pid_max
)
587 set_bit(pid
, pid_list
->pids
);
590 trace_parser_clear(&parser
);
593 trace_parser_put(&parser
);
596 trace_free_pid_list(pid_list
);
601 /* Cleared the list of pids */
602 trace_free_pid_list(pid_list
);
607 *new_pid_list
= pid_list
;
612 static u64
buffer_ftrace_now(struct array_buffer
*buf
, int cpu
)
616 /* Early boot up does not have a buffer yet */
618 return trace_clock_local();
620 ts
= ring_buffer_time_stamp(buf
->buffer
, cpu
);
621 ring_buffer_normalize_time_stamp(buf
->buffer
, cpu
, &ts
);
626 u64
ftrace_now(int cpu
)
628 return buffer_ftrace_now(&global_trace
.array_buffer
, cpu
);
632 * tracing_is_enabled - Show if global_trace has been disabled
634 * Shows if the global trace has been enabled or not. It uses the
635 * mirror flag "buffer_disabled" to be used in fast paths such as for
636 * the irqsoff tracer. But it may be inaccurate due to races. If you
637 * need to know the accurate state, use tracing_is_on() which is a little
638 * slower, but accurate.
640 int tracing_is_enabled(void)
643 * For quick access (irqsoff uses this in fast path), just
644 * return the mirror variable of the state of the ring buffer.
645 * It's a little racy, but we don't really care.
648 return !global_trace
.buffer_disabled
;
652 * trace_buf_size is the size in bytes that is allocated
653 * for a buffer. Note, the number of bytes is always rounded
656 * This number is purposely set to a low number of 16384.
657 * If the dump on oops happens, it will be much appreciated
658 * to not have to wait for all that output. Anyway this can be
659 * boot time and run time configurable.
661 #define TRACE_BUF_SIZE_DEFAULT 1441792UL /* 16384 * 88 (sizeof(entry)) */
663 static unsigned long trace_buf_size
= TRACE_BUF_SIZE_DEFAULT
;
665 /* trace_types holds a link list of available tracers. */
666 static struct tracer
*trace_types __read_mostly
;
669 * trace_types_lock is used to protect the trace_types list.
671 DEFINE_MUTEX(trace_types_lock
);
674 * serialize the access of the ring buffer
676 * ring buffer serializes readers, but it is low level protection.
677 * The validity of the events (which returns by ring_buffer_peek() ..etc)
678 * are not protected by ring buffer.
680 * The content of events may become garbage if we allow other process consumes
681 * these events concurrently:
682 * A) the page of the consumed events may become a normal page
683 * (not reader page) in ring buffer, and this page will be rewrited
684 * by events producer.
685 * B) The page of the consumed events may become a page for splice_read,
686 * and this page will be returned to system.
688 * These primitives allow multi process access to different cpu ring buffer
691 * These primitives don't distinguish read-only and read-consume access.
692 * Multi read-only access are also serialized.
696 static DECLARE_RWSEM(all_cpu_access_lock
);
697 static DEFINE_PER_CPU(struct mutex
, cpu_access_lock
);
699 static inline void trace_access_lock(int cpu
)
701 if (cpu
== RING_BUFFER_ALL_CPUS
) {
702 /* gain it for accessing the whole ring buffer. */
703 down_write(&all_cpu_access_lock
);
705 /* gain it for accessing a cpu ring buffer. */
707 /* Firstly block other trace_access_lock(RING_BUFFER_ALL_CPUS). */
708 down_read(&all_cpu_access_lock
);
710 /* Secondly block other access to this @cpu ring buffer. */
711 mutex_lock(&per_cpu(cpu_access_lock
, cpu
));
715 static inline void trace_access_unlock(int cpu
)
717 if (cpu
== RING_BUFFER_ALL_CPUS
) {
718 up_write(&all_cpu_access_lock
);
720 mutex_unlock(&per_cpu(cpu_access_lock
, cpu
));
721 up_read(&all_cpu_access_lock
);
725 static inline void trace_access_lock_init(void)
729 for_each_possible_cpu(cpu
)
730 mutex_init(&per_cpu(cpu_access_lock
, cpu
));
735 static DEFINE_MUTEX(access_lock
);
737 static inline void trace_access_lock(int cpu
)
740 mutex_lock(&access_lock
);
743 static inline void trace_access_unlock(int cpu
)
746 mutex_unlock(&access_lock
);
749 static inline void trace_access_lock_init(void)
755 #ifdef CONFIG_STACKTRACE
756 static void __ftrace_trace_stack(struct trace_buffer
*buffer
,
758 int skip
, int pc
, struct pt_regs
*regs
);
759 static inline void ftrace_trace_stack(struct trace_array
*tr
,
760 struct trace_buffer
*buffer
,
762 int skip
, int pc
, struct pt_regs
*regs
);
765 static inline void __ftrace_trace_stack(struct trace_buffer
*buffer
,
767 int skip
, int pc
, struct pt_regs
*regs
)
770 static inline void ftrace_trace_stack(struct trace_array
*tr
,
771 struct trace_buffer
*buffer
,
773 int skip
, int pc
, struct pt_regs
*regs
)
779 static __always_inline
void
780 trace_event_setup(struct ring_buffer_event
*event
,
781 int type
, unsigned long flags
, int pc
)
783 struct trace_entry
*ent
= ring_buffer_event_data(event
);
785 tracing_generic_entry_update(ent
, type
, flags
, pc
);
788 static __always_inline
struct ring_buffer_event
*
789 __trace_buffer_lock_reserve(struct trace_buffer
*buffer
,
792 unsigned long flags
, int pc
)
794 struct ring_buffer_event
*event
;
796 event
= ring_buffer_lock_reserve(buffer
, len
);
798 trace_event_setup(event
, type
, flags
, pc
);
803 void tracer_tracing_on(struct trace_array
*tr
)
805 if (tr
->array_buffer
.buffer
)
806 ring_buffer_record_on(tr
->array_buffer
.buffer
);
808 * This flag is looked at when buffers haven't been allocated
809 * yet, or by some tracers (like irqsoff), that just want to
810 * know if the ring buffer has been disabled, but it can handle
811 * races of where it gets disabled but we still do a record.
812 * As the check is in the fast path of the tracers, it is more
813 * important to be fast than accurate.
815 tr
->buffer_disabled
= 0;
816 /* Make the flag seen by readers */
821 * tracing_on - enable tracing buffers
823 * This function enables tracing buffers that may have been
824 * disabled with tracing_off.
826 void tracing_on(void)
828 tracer_tracing_on(&global_trace
);
830 EXPORT_SYMBOL_GPL(tracing_on
);
833 static __always_inline
void
834 __buffer_unlock_commit(struct trace_buffer
*buffer
, struct ring_buffer_event
*event
)
836 __this_cpu_write(trace_taskinfo_save
, true);
838 /* If this is the temp buffer, we need to commit fully */
839 if (this_cpu_read(trace_buffered_event
) == event
) {
840 /* Length is in event->array[0] */
841 ring_buffer_write(buffer
, event
->array
[0], &event
->array
[1]);
842 /* Release the temp buffer */
843 this_cpu_dec(trace_buffered_event_cnt
);
845 ring_buffer_unlock_commit(buffer
, event
);
849 * __trace_puts - write a constant string into the trace buffer.
850 * @ip: The address of the caller
851 * @str: The constant string to write
852 * @size: The size of the string.
854 int __trace_puts(unsigned long ip
, const char *str
, int size
)
856 struct ring_buffer_event
*event
;
857 struct trace_buffer
*buffer
;
858 struct print_entry
*entry
;
859 unsigned long irq_flags
;
863 if (!(global_trace
.trace_flags
& TRACE_ITER_PRINTK
))
866 pc
= preempt_count();
868 if (unlikely(tracing_selftest_running
|| tracing_disabled
))
871 alloc
= sizeof(*entry
) + size
+ 2; /* possible \n added */
873 local_save_flags(irq_flags
);
874 buffer
= global_trace
.array_buffer
.buffer
;
875 ring_buffer_nest_start(buffer
);
876 event
= __trace_buffer_lock_reserve(buffer
, TRACE_PRINT
, alloc
,
883 entry
= ring_buffer_event_data(event
);
886 memcpy(&entry
->buf
, str
, size
);
888 /* Add a newline if necessary */
889 if (entry
->buf
[size
- 1] != '\n') {
890 entry
->buf
[size
] = '\n';
891 entry
->buf
[size
+ 1] = '\0';
893 entry
->buf
[size
] = '\0';
895 __buffer_unlock_commit(buffer
, event
);
896 ftrace_trace_stack(&global_trace
, buffer
, irq_flags
, 4, pc
, NULL
);
898 ring_buffer_nest_end(buffer
);
901 EXPORT_SYMBOL_GPL(__trace_puts
);
904 * __trace_bputs - write the pointer to a constant string into trace buffer
905 * @ip: The address of the caller
906 * @str: The constant string to write to the buffer to
908 int __trace_bputs(unsigned long ip
, const char *str
)
910 struct ring_buffer_event
*event
;
911 struct trace_buffer
*buffer
;
912 struct bputs_entry
*entry
;
913 unsigned long irq_flags
;
914 int size
= sizeof(struct bputs_entry
);
918 if (!(global_trace
.trace_flags
& TRACE_ITER_PRINTK
))
921 pc
= preempt_count();
923 if (unlikely(tracing_selftest_running
|| tracing_disabled
))
926 local_save_flags(irq_flags
);
927 buffer
= global_trace
.array_buffer
.buffer
;
929 ring_buffer_nest_start(buffer
);
930 event
= __trace_buffer_lock_reserve(buffer
, TRACE_BPUTS
, size
,
935 entry
= ring_buffer_event_data(event
);
939 __buffer_unlock_commit(buffer
, event
);
940 ftrace_trace_stack(&global_trace
, buffer
, irq_flags
, 4, pc
, NULL
);
944 ring_buffer_nest_end(buffer
);
947 EXPORT_SYMBOL_GPL(__trace_bputs
);
949 #ifdef CONFIG_TRACER_SNAPSHOT
950 static void tracing_snapshot_instance_cond(struct trace_array
*tr
,
953 struct tracer
*tracer
= tr
->current_trace
;
957 internal_trace_puts("*** SNAPSHOT CALLED FROM NMI CONTEXT ***\n");
958 internal_trace_puts("*** snapshot is being ignored ***\n");
962 if (!tr
->allocated_snapshot
) {
963 internal_trace_puts("*** SNAPSHOT NOT ALLOCATED ***\n");
964 internal_trace_puts("*** stopping trace here! ***\n");
969 /* Note, snapshot can not be used when the tracer uses it */
970 if (tracer
->use_max_tr
) {
971 internal_trace_puts("*** LATENCY TRACER ACTIVE ***\n");
972 internal_trace_puts("*** Can not use snapshot (sorry) ***\n");
976 local_irq_save(flags
);
977 update_max_tr(tr
, current
, smp_processor_id(), cond_data
);
978 local_irq_restore(flags
);
981 void tracing_snapshot_instance(struct trace_array
*tr
)
983 tracing_snapshot_instance_cond(tr
, NULL
);
987 * tracing_snapshot - take a snapshot of the current buffer.
989 * This causes a swap between the snapshot buffer and the current live
990 * tracing buffer. You can use this to take snapshots of the live
991 * trace when some condition is triggered, but continue to trace.
993 * Note, make sure to allocate the snapshot with either
994 * a tracing_snapshot_alloc(), or by doing it manually
995 * with: echo 1 > /sys/kernel/debug/tracing/snapshot
997 * If the snapshot buffer is not allocated, it will stop tracing.
998 * Basically making a permanent snapshot.
1000 void tracing_snapshot(void)
1002 struct trace_array
*tr
= &global_trace
;
1004 tracing_snapshot_instance(tr
);
1006 EXPORT_SYMBOL_GPL(tracing_snapshot
);
1009 * tracing_snapshot_cond - conditionally take a snapshot of the current buffer.
1010 * @tr: The tracing instance to snapshot
1011 * @cond_data: The data to be tested conditionally, and possibly saved
1013 * This is the same as tracing_snapshot() except that the snapshot is
1014 * conditional - the snapshot will only happen if the
1015 * cond_snapshot.update() implementation receiving the cond_data
1016 * returns true, which means that the trace array's cond_snapshot
1017 * update() operation used the cond_data to determine whether the
1018 * snapshot should be taken, and if it was, presumably saved it along
1019 * with the snapshot.
1021 void tracing_snapshot_cond(struct trace_array
*tr
, void *cond_data
)
1023 tracing_snapshot_instance_cond(tr
, cond_data
);
1025 EXPORT_SYMBOL_GPL(tracing_snapshot_cond
);
1028 * tracing_snapshot_cond_data - get the user data associated with a snapshot
1029 * @tr: The tracing instance
1031 * When the user enables a conditional snapshot using
1032 * tracing_snapshot_cond_enable(), the user-defined cond_data is saved
1033 * with the snapshot. This accessor is used to retrieve it.
1035 * Should not be called from cond_snapshot.update(), since it takes
1036 * the tr->max_lock lock, which the code calling
1037 * cond_snapshot.update() has already done.
1039 * Returns the cond_data associated with the trace array's snapshot.
1041 void *tracing_cond_snapshot_data(struct trace_array
*tr
)
1043 void *cond_data
= NULL
;
1045 arch_spin_lock(&tr
->max_lock
);
1047 if (tr
->cond_snapshot
)
1048 cond_data
= tr
->cond_snapshot
->cond_data
;
1050 arch_spin_unlock(&tr
->max_lock
);
1054 EXPORT_SYMBOL_GPL(tracing_cond_snapshot_data
);
1056 static int resize_buffer_duplicate_size(struct array_buffer
*trace_buf
,
1057 struct array_buffer
*size_buf
, int cpu_id
);
1058 static void set_buffer_entries(struct array_buffer
*buf
, unsigned long val
);
1060 int tracing_alloc_snapshot_instance(struct trace_array
*tr
)
1064 if (!tr
->allocated_snapshot
) {
1066 /* allocate spare buffer */
1067 ret
= resize_buffer_duplicate_size(&tr
->max_buffer
,
1068 &tr
->array_buffer
, RING_BUFFER_ALL_CPUS
);
1072 tr
->allocated_snapshot
= true;
1078 static void free_snapshot(struct trace_array
*tr
)
1081 * We don't free the ring buffer. instead, resize it because
1082 * The max_tr ring buffer has some state (e.g. ring->clock) and
1083 * we want preserve it.
1085 ring_buffer_resize(tr
->max_buffer
.buffer
, 1, RING_BUFFER_ALL_CPUS
);
1086 set_buffer_entries(&tr
->max_buffer
, 1);
1087 tracing_reset_online_cpus(&tr
->max_buffer
);
1088 tr
->allocated_snapshot
= false;
1092 * tracing_alloc_snapshot - allocate snapshot buffer.
1094 * This only allocates the snapshot buffer if it isn't already
1095 * allocated - it doesn't also take a snapshot.
1097 * This is meant to be used in cases where the snapshot buffer needs
1098 * to be set up for events that can't sleep but need to be able to
1099 * trigger a snapshot.
1101 int tracing_alloc_snapshot(void)
1103 struct trace_array
*tr
= &global_trace
;
1106 ret
= tracing_alloc_snapshot_instance(tr
);
1111 EXPORT_SYMBOL_GPL(tracing_alloc_snapshot
);
1114 * tracing_snapshot_alloc - allocate and take a snapshot of the current buffer.
1116 * This is similar to tracing_snapshot(), but it will allocate the
1117 * snapshot buffer if it isn't already allocated. Use this only
1118 * where it is safe to sleep, as the allocation may sleep.
1120 * This causes a swap between the snapshot buffer and the current live
1121 * tracing buffer. You can use this to take snapshots of the live
1122 * trace when some condition is triggered, but continue to trace.
1124 void tracing_snapshot_alloc(void)
1128 ret
= tracing_alloc_snapshot();
1134 EXPORT_SYMBOL_GPL(tracing_snapshot_alloc
);
1137 * tracing_snapshot_cond_enable - enable conditional snapshot for an instance
1138 * @tr: The tracing instance
1139 * @cond_data: User data to associate with the snapshot
1140 * @update: Implementation of the cond_snapshot update function
1142 * Check whether the conditional snapshot for the given instance has
1143 * already been enabled, or if the current tracer is already using a
1144 * snapshot; if so, return -EBUSY, else create a cond_snapshot and
1145 * save the cond_data and update function inside.
1147 * Returns 0 if successful, error otherwise.
1149 int tracing_snapshot_cond_enable(struct trace_array
*tr
, void *cond_data
,
1150 cond_update_fn_t update
)
1152 struct cond_snapshot
*cond_snapshot
;
1155 cond_snapshot
= kzalloc(sizeof(*cond_snapshot
), GFP_KERNEL
);
1159 cond_snapshot
->cond_data
= cond_data
;
1160 cond_snapshot
->update
= update
;
1162 mutex_lock(&trace_types_lock
);
1164 ret
= tracing_alloc_snapshot_instance(tr
);
1168 if (tr
->current_trace
->use_max_tr
) {
1174 * The cond_snapshot can only change to NULL without the
1175 * trace_types_lock. We don't care if we race with it going
1176 * to NULL, but we want to make sure that it's not set to
1177 * something other than NULL when we get here, which we can
1178 * do safely with only holding the trace_types_lock and not
1179 * having to take the max_lock.
1181 if (tr
->cond_snapshot
) {
1186 arch_spin_lock(&tr
->max_lock
);
1187 tr
->cond_snapshot
= cond_snapshot
;
1188 arch_spin_unlock(&tr
->max_lock
);
1190 mutex_unlock(&trace_types_lock
);
1195 mutex_unlock(&trace_types_lock
);
1196 kfree(cond_snapshot
);
1199 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_enable
);
1202 * tracing_snapshot_cond_disable - disable conditional snapshot for an instance
1203 * @tr: The tracing instance
1205 * Check whether the conditional snapshot for the given instance is
1206 * enabled; if so, free the cond_snapshot associated with it,
1207 * otherwise return -EINVAL.
1209 * Returns 0 if successful, error otherwise.
1211 int tracing_snapshot_cond_disable(struct trace_array
*tr
)
1215 arch_spin_lock(&tr
->max_lock
);
1217 if (!tr
->cond_snapshot
)
1220 kfree(tr
->cond_snapshot
);
1221 tr
->cond_snapshot
= NULL
;
1224 arch_spin_unlock(&tr
->max_lock
);
1228 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_disable
);
1230 void tracing_snapshot(void)
1232 WARN_ONCE(1, "Snapshot feature not enabled, but internal snapshot used");
1234 EXPORT_SYMBOL_GPL(tracing_snapshot
);
1235 void tracing_snapshot_cond(struct trace_array
*tr
, void *cond_data
)
1237 WARN_ONCE(1, "Snapshot feature not enabled, but internal conditional snapshot used");
1239 EXPORT_SYMBOL_GPL(tracing_snapshot_cond
);
1240 int tracing_alloc_snapshot(void)
1242 WARN_ONCE(1, "Snapshot feature not enabled, but snapshot allocation used");
1245 EXPORT_SYMBOL_GPL(tracing_alloc_snapshot
);
1246 void tracing_snapshot_alloc(void)
1251 EXPORT_SYMBOL_GPL(tracing_snapshot_alloc
);
1252 void *tracing_cond_snapshot_data(struct trace_array
*tr
)
1256 EXPORT_SYMBOL_GPL(tracing_cond_snapshot_data
);
1257 int tracing_snapshot_cond_enable(struct trace_array
*tr
, void *cond_data
, cond_update_fn_t update
)
1261 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_enable
);
1262 int tracing_snapshot_cond_disable(struct trace_array
*tr
)
1266 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_disable
);
1267 #endif /* CONFIG_TRACER_SNAPSHOT */
1269 void tracer_tracing_off(struct trace_array
*tr
)
1271 if (tr
->array_buffer
.buffer
)
1272 ring_buffer_record_off(tr
->array_buffer
.buffer
);
1274 * This flag is looked at when buffers haven't been allocated
1275 * yet, or by some tracers (like irqsoff), that just want to
1276 * know if the ring buffer has been disabled, but it can handle
1277 * races of where it gets disabled but we still do a record.
1278 * As the check is in the fast path of the tracers, it is more
1279 * important to be fast than accurate.
1281 tr
->buffer_disabled
= 1;
1282 /* Make the flag seen by readers */
1287 * tracing_off - turn off tracing buffers
1289 * This function stops the tracing buffers from recording data.
1290 * It does not disable any overhead the tracers themselves may
1291 * be causing. This function simply causes all recording to
1292 * the ring buffers to fail.
1294 void tracing_off(void)
1296 tracer_tracing_off(&global_trace
);
1298 EXPORT_SYMBOL_GPL(tracing_off
);
1300 void disable_trace_on_warning(void)
1302 if (__disable_trace_on_warning
) {
1303 trace_array_printk_buf(global_trace
.array_buffer
.buffer
, _THIS_IP_
,
1304 "Disabling tracing due to warning\n");
1310 * tracer_tracing_is_on - show real state of ring buffer enabled
1311 * @tr : the trace array to know if ring buffer is enabled
1313 * Shows real state of the ring buffer if it is enabled or not.
1315 bool tracer_tracing_is_on(struct trace_array
*tr
)
1317 if (tr
->array_buffer
.buffer
)
1318 return ring_buffer_record_is_on(tr
->array_buffer
.buffer
);
1319 return !tr
->buffer_disabled
;
1323 * tracing_is_on - show state of ring buffers enabled
1325 int tracing_is_on(void)
1327 return tracer_tracing_is_on(&global_trace
);
1329 EXPORT_SYMBOL_GPL(tracing_is_on
);
1331 static int __init
set_buf_size(char *str
)
1333 unsigned long buf_size
;
1337 buf_size
= memparse(str
, &str
);
1338 /* nr_entries can not be zero */
1341 trace_buf_size
= buf_size
;
1344 __setup("trace_buf_size=", set_buf_size
);
1346 static int __init
set_tracing_thresh(char *str
)
1348 unsigned long threshold
;
1353 ret
= kstrtoul(str
, 0, &threshold
);
1356 tracing_thresh
= threshold
* 1000;
1359 __setup("tracing_thresh=", set_tracing_thresh
);
1361 unsigned long nsecs_to_usecs(unsigned long nsecs
)
1363 return nsecs
/ 1000;
1367 * TRACE_FLAGS is defined as a tuple matching bit masks with strings.
1368 * It uses C(a, b) where 'a' is the eval (enum) name and 'b' is the string that
1369 * matches it. By defining "C(a, b) b", TRACE_FLAGS becomes a list
1370 * of strings in the order that the evals (enum) were defined.
1375 /* These must match the bit postions in trace_iterator_flags */
1376 static const char *trace_options
[] = {
1384 int in_ns
; /* is this clock in nanoseconds? */
1385 } trace_clocks
[] = {
1386 { trace_clock_local
, "local", 1 },
1387 { trace_clock_global
, "global", 1 },
1388 { trace_clock_counter
, "counter", 0 },
1389 { trace_clock_jiffies
, "uptime", 0 },
1390 { trace_clock
, "perf", 1 },
1391 { ktime_get_mono_fast_ns
, "mono", 1 },
1392 { ktime_get_raw_fast_ns
, "mono_raw", 1 },
1393 { ktime_get_boot_fast_ns
, "boot", 1 },
1397 bool trace_clock_in_ns(struct trace_array
*tr
)
1399 if (trace_clocks
[tr
->clock_id
].in_ns
)
1406 * trace_parser_get_init - gets the buffer for trace parser
1408 int trace_parser_get_init(struct trace_parser
*parser
, int size
)
1410 memset(parser
, 0, sizeof(*parser
));
1412 parser
->buffer
= kmalloc(size
, GFP_KERNEL
);
1413 if (!parser
->buffer
)
1416 parser
->size
= size
;
1421 * trace_parser_put - frees the buffer for trace parser
1423 void trace_parser_put(struct trace_parser
*parser
)
1425 kfree(parser
->buffer
);
1426 parser
->buffer
= NULL
;
1430 * trace_get_user - reads the user input string separated by space
1431 * (matched by isspace(ch))
1433 * For each string found the 'struct trace_parser' is updated,
1434 * and the function returns.
1436 * Returns number of bytes read.
1438 * See kernel/trace/trace.h for 'struct trace_parser' details.
1440 int trace_get_user(struct trace_parser
*parser
, const char __user
*ubuf
,
1441 size_t cnt
, loff_t
*ppos
)
1448 trace_parser_clear(parser
);
1450 ret
= get_user(ch
, ubuf
++);
1458 * The parser is not finished with the last write,
1459 * continue reading the user input without skipping spaces.
1461 if (!parser
->cont
) {
1462 /* skip white space */
1463 while (cnt
&& isspace(ch
)) {
1464 ret
= get_user(ch
, ubuf
++);
1473 /* only spaces were written */
1474 if (isspace(ch
) || !ch
) {
1481 /* read the non-space input */
1482 while (cnt
&& !isspace(ch
) && ch
) {
1483 if (parser
->idx
< parser
->size
- 1)
1484 parser
->buffer
[parser
->idx
++] = ch
;
1489 ret
= get_user(ch
, ubuf
++);
1496 /* We either got finished input or we have to wait for another call. */
1497 if (isspace(ch
) || !ch
) {
1498 parser
->buffer
[parser
->idx
] = 0;
1499 parser
->cont
= false;
1500 } else if (parser
->idx
< parser
->size
- 1) {
1501 parser
->cont
= true;
1502 parser
->buffer
[parser
->idx
++] = ch
;
1503 /* Make sure the parsed string always terminates with '\0'. */
1504 parser
->buffer
[parser
->idx
] = 0;
1517 /* TODO add a seq_buf_to_buffer() */
1518 static ssize_t
trace_seq_to_buffer(struct trace_seq
*s
, void *buf
, size_t cnt
)
1522 if (trace_seq_used(s
) <= s
->seq
.readpos
)
1525 len
= trace_seq_used(s
) - s
->seq
.readpos
;
1528 memcpy(buf
, s
->buffer
+ s
->seq
.readpos
, cnt
);
1530 s
->seq
.readpos
+= cnt
;
1534 unsigned long __read_mostly tracing_thresh
;
1535 static const struct file_operations tracing_max_lat_fops
;
1537 #if (defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)) && \
1538 defined(CONFIG_FSNOTIFY)
1540 static struct workqueue_struct
*fsnotify_wq
;
1542 static void latency_fsnotify_workfn(struct work_struct
*work
)
1544 struct trace_array
*tr
= container_of(work
, struct trace_array
,
1546 fsnotify(tr
->d_max_latency
->d_inode
, FS_MODIFY
,
1547 tr
->d_max_latency
->d_inode
, FSNOTIFY_EVENT_INODE
, NULL
, 0);
1550 static void latency_fsnotify_workfn_irq(struct irq_work
*iwork
)
1552 struct trace_array
*tr
= container_of(iwork
, struct trace_array
,
1554 queue_work(fsnotify_wq
, &tr
->fsnotify_work
);
1557 static void trace_create_maxlat_file(struct trace_array
*tr
,
1558 struct dentry
*d_tracer
)
1560 INIT_WORK(&tr
->fsnotify_work
, latency_fsnotify_workfn
);
1561 init_irq_work(&tr
->fsnotify_irqwork
, latency_fsnotify_workfn_irq
);
1562 tr
->d_max_latency
= trace_create_file("tracing_max_latency", 0644,
1563 d_tracer
, &tr
->max_latency
,
1564 &tracing_max_lat_fops
);
1567 __init
static int latency_fsnotify_init(void)
1569 fsnotify_wq
= alloc_workqueue("tr_max_lat_wq",
1570 WQ_UNBOUND
| WQ_HIGHPRI
, 0);
1572 pr_err("Unable to allocate tr_max_lat_wq\n");
1578 late_initcall_sync(latency_fsnotify_init
);
1580 void latency_fsnotify(struct trace_array
*tr
)
1585 * We cannot call queue_work(&tr->fsnotify_work) from here because it's
1586 * possible that we are called from __schedule() or do_idle(), which
1587 * could cause a deadlock.
1589 irq_work_queue(&tr
->fsnotify_irqwork
);
1593 * (defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)) && \
1594 * defined(CONFIG_FSNOTIFY)
1598 #define trace_create_maxlat_file(tr, d_tracer) \
1599 trace_create_file("tracing_max_latency", 0644, d_tracer, \
1600 &tr->max_latency, &tracing_max_lat_fops)
1604 #ifdef CONFIG_TRACER_MAX_TRACE
1606 * Copy the new maximum trace into the separate maximum-trace
1607 * structure. (this way the maximum trace is permanently saved,
1608 * for later retrieval via /sys/kernel/tracing/tracing_max_latency)
1611 __update_max_tr(struct trace_array
*tr
, struct task_struct
*tsk
, int cpu
)
1613 struct array_buffer
*trace_buf
= &tr
->array_buffer
;
1614 struct array_buffer
*max_buf
= &tr
->max_buffer
;
1615 struct trace_array_cpu
*data
= per_cpu_ptr(trace_buf
->data
, cpu
);
1616 struct trace_array_cpu
*max_data
= per_cpu_ptr(max_buf
->data
, cpu
);
1619 max_buf
->time_start
= data
->preempt_timestamp
;
1621 max_data
->saved_latency
= tr
->max_latency
;
1622 max_data
->critical_start
= data
->critical_start
;
1623 max_data
->critical_end
= data
->critical_end
;
1625 strncpy(max_data
->comm
, tsk
->comm
, TASK_COMM_LEN
);
1626 max_data
->pid
= tsk
->pid
;
1628 * If tsk == current, then use current_uid(), as that does not use
1629 * RCU. The irq tracer can be called out of RCU scope.
1632 max_data
->uid
= current_uid();
1634 max_data
->uid
= task_uid(tsk
);
1636 max_data
->nice
= tsk
->static_prio
- 20 - MAX_RT_PRIO
;
1637 max_data
->policy
= tsk
->policy
;
1638 max_data
->rt_priority
= tsk
->rt_priority
;
1640 /* record this tasks comm */
1641 tracing_record_cmdline(tsk
);
1642 latency_fsnotify(tr
);
1646 * update_max_tr - snapshot all trace buffers from global_trace to max_tr
1648 * @tsk: the task with the latency
1649 * @cpu: The cpu that initiated the trace.
1650 * @cond_data: User data associated with a conditional snapshot
1652 * Flip the buffers between the @tr and the max_tr and record information
1653 * about which task was the cause of this latency.
1656 update_max_tr(struct trace_array
*tr
, struct task_struct
*tsk
, int cpu
,
1662 WARN_ON_ONCE(!irqs_disabled());
1664 if (!tr
->allocated_snapshot
) {
1665 /* Only the nop tracer should hit this when disabling */
1666 WARN_ON_ONCE(tr
->current_trace
!= &nop_trace
);
1670 arch_spin_lock(&tr
->max_lock
);
1672 /* Inherit the recordable setting from array_buffer */
1673 if (ring_buffer_record_is_set_on(tr
->array_buffer
.buffer
))
1674 ring_buffer_record_on(tr
->max_buffer
.buffer
);
1676 ring_buffer_record_off(tr
->max_buffer
.buffer
);
1678 #ifdef CONFIG_TRACER_SNAPSHOT
1679 if (tr
->cond_snapshot
&& !tr
->cond_snapshot
->update(tr
, cond_data
))
1682 swap(tr
->array_buffer
.buffer
, tr
->max_buffer
.buffer
);
1684 __update_max_tr(tr
, tsk
, cpu
);
1687 arch_spin_unlock(&tr
->max_lock
);
1691 * update_max_tr_single - only copy one trace over, and reset the rest
1693 * @tsk: task with the latency
1694 * @cpu: the cpu of the buffer to copy.
1696 * Flip the trace of a single CPU buffer between the @tr and the max_tr.
1699 update_max_tr_single(struct trace_array
*tr
, struct task_struct
*tsk
, int cpu
)
1706 WARN_ON_ONCE(!irqs_disabled());
1707 if (!tr
->allocated_snapshot
) {
1708 /* Only the nop tracer should hit this when disabling */
1709 WARN_ON_ONCE(tr
->current_trace
!= &nop_trace
);
1713 arch_spin_lock(&tr
->max_lock
);
1715 ret
= ring_buffer_swap_cpu(tr
->max_buffer
.buffer
, tr
->array_buffer
.buffer
, cpu
);
1717 if (ret
== -EBUSY
) {
1719 * We failed to swap the buffer due to a commit taking
1720 * place on this CPU. We fail to record, but we reset
1721 * the max trace buffer (no one writes directly to it)
1722 * and flag that it failed.
1724 trace_array_printk_buf(tr
->max_buffer
.buffer
, _THIS_IP_
,
1725 "Failed to swap buffers due to commit in progress\n");
1728 WARN_ON_ONCE(ret
&& ret
!= -EAGAIN
&& ret
!= -EBUSY
);
1730 __update_max_tr(tr
, tsk
, cpu
);
1731 arch_spin_unlock(&tr
->max_lock
);
1733 #endif /* CONFIG_TRACER_MAX_TRACE */
1735 static int wait_on_pipe(struct trace_iterator
*iter
, int full
)
1737 /* Iterators are static, they should be filled or empty */
1738 if (trace_buffer_iter(iter
, iter
->cpu_file
))
1741 return ring_buffer_wait(iter
->array_buffer
->buffer
, iter
->cpu_file
,
1745 #ifdef CONFIG_FTRACE_STARTUP_TEST
1746 static bool selftests_can_run
;
1748 struct trace_selftests
{
1749 struct list_head list
;
1750 struct tracer
*type
;
1753 static LIST_HEAD(postponed_selftests
);
1755 static int save_selftest(struct tracer
*type
)
1757 struct trace_selftests
*selftest
;
1759 selftest
= kmalloc(sizeof(*selftest
), GFP_KERNEL
);
1763 selftest
->type
= type
;
1764 list_add(&selftest
->list
, &postponed_selftests
);
1768 static int run_tracer_selftest(struct tracer
*type
)
1770 struct trace_array
*tr
= &global_trace
;
1771 struct tracer
*saved_tracer
= tr
->current_trace
;
1774 if (!type
->selftest
|| tracing_selftest_disabled
)
1778 * If a tracer registers early in boot up (before scheduling is
1779 * initialized and such), then do not run its selftests yet.
1780 * Instead, run it a little later in the boot process.
1782 if (!selftests_can_run
)
1783 return save_selftest(type
);
1786 * Run a selftest on this tracer.
1787 * Here we reset the trace buffer, and set the current
1788 * tracer to be this tracer. The tracer can then run some
1789 * internal tracing to verify that everything is in order.
1790 * If we fail, we do not register this tracer.
1792 tracing_reset_online_cpus(&tr
->array_buffer
);
1794 tr
->current_trace
= type
;
1796 #ifdef CONFIG_TRACER_MAX_TRACE
1797 if (type
->use_max_tr
) {
1798 /* If we expanded the buffers, make sure the max is expanded too */
1799 if (ring_buffer_expanded
)
1800 ring_buffer_resize(tr
->max_buffer
.buffer
, trace_buf_size
,
1801 RING_BUFFER_ALL_CPUS
);
1802 tr
->allocated_snapshot
= true;
1806 /* the test is responsible for initializing and enabling */
1807 pr_info("Testing tracer %s: ", type
->name
);
1808 ret
= type
->selftest(type
, tr
);
1809 /* the test is responsible for resetting too */
1810 tr
->current_trace
= saved_tracer
;
1812 printk(KERN_CONT
"FAILED!\n");
1813 /* Add the warning after printing 'FAILED' */
1817 /* Only reset on passing, to avoid touching corrupted buffers */
1818 tracing_reset_online_cpus(&tr
->array_buffer
);
1820 #ifdef CONFIG_TRACER_MAX_TRACE
1821 if (type
->use_max_tr
) {
1822 tr
->allocated_snapshot
= false;
1824 /* Shrink the max buffer again */
1825 if (ring_buffer_expanded
)
1826 ring_buffer_resize(tr
->max_buffer
.buffer
, 1,
1827 RING_BUFFER_ALL_CPUS
);
1831 printk(KERN_CONT
"PASSED\n");
1835 static __init
int init_trace_selftests(void)
1837 struct trace_selftests
*p
, *n
;
1838 struct tracer
*t
, **last
;
1841 selftests_can_run
= true;
1843 mutex_lock(&trace_types_lock
);
1845 if (list_empty(&postponed_selftests
))
1848 pr_info("Running postponed tracer tests:\n");
1850 tracing_selftest_running
= true;
1851 list_for_each_entry_safe(p
, n
, &postponed_selftests
, list
) {
1852 /* This loop can take minutes when sanitizers are enabled, so
1853 * lets make sure we allow RCU processing.
1856 ret
= run_tracer_selftest(p
->type
);
1857 /* If the test fails, then warn and remove from available_tracers */
1859 WARN(1, "tracer: %s failed selftest, disabling\n",
1861 last
= &trace_types
;
1862 for (t
= trace_types
; t
; t
= t
->next
) {
1873 tracing_selftest_running
= false;
1876 mutex_unlock(&trace_types_lock
);
1880 core_initcall(init_trace_selftests
);
1882 static inline int run_tracer_selftest(struct tracer
*type
)
1886 #endif /* CONFIG_FTRACE_STARTUP_TEST */
1888 static void add_tracer_options(struct trace_array
*tr
, struct tracer
*t
);
1890 static void __init
apply_trace_boot_options(void);
1893 * register_tracer - register a tracer with the ftrace system.
1894 * @type: the plugin for the tracer
1896 * Register a new plugin tracer.
1898 int __init
register_tracer(struct tracer
*type
)
1904 pr_info("Tracer must have a name\n");
1908 if (strlen(type
->name
) >= MAX_TRACER_SIZE
) {
1909 pr_info("Tracer has a name longer than %d\n", MAX_TRACER_SIZE
);
1913 if (security_locked_down(LOCKDOWN_TRACEFS
)) {
1914 pr_warn("Can not register tracer %s due to lockdown\n",
1919 mutex_lock(&trace_types_lock
);
1921 tracing_selftest_running
= true;
1923 for (t
= trace_types
; t
; t
= t
->next
) {
1924 if (strcmp(type
->name
, t
->name
) == 0) {
1926 pr_info("Tracer %s already registered\n",
1933 if (!type
->set_flag
)
1934 type
->set_flag
= &dummy_set_flag
;
1936 /*allocate a dummy tracer_flags*/
1937 type
->flags
= kmalloc(sizeof(*type
->flags
), GFP_KERNEL
);
1942 type
->flags
->val
= 0;
1943 type
->flags
->opts
= dummy_tracer_opt
;
1945 if (!type
->flags
->opts
)
1946 type
->flags
->opts
= dummy_tracer_opt
;
1948 /* store the tracer for __set_tracer_option */
1949 type
->flags
->trace
= type
;
1951 ret
= run_tracer_selftest(type
);
1955 type
->next
= trace_types
;
1957 add_tracer_options(&global_trace
, type
);
1960 tracing_selftest_running
= false;
1961 mutex_unlock(&trace_types_lock
);
1963 if (ret
|| !default_bootup_tracer
)
1966 if (strncmp(default_bootup_tracer
, type
->name
, MAX_TRACER_SIZE
))
1969 printk(KERN_INFO
"Starting tracer '%s'\n", type
->name
);
1970 /* Do we want this tracer to start on bootup? */
1971 tracing_set_tracer(&global_trace
, type
->name
);
1972 default_bootup_tracer
= NULL
;
1974 apply_trace_boot_options();
1976 /* disable other selftests, since this will break it. */
1977 tracing_selftest_disabled
= true;
1978 #ifdef CONFIG_FTRACE_STARTUP_TEST
1979 printk(KERN_INFO
"Disabling FTRACE selftests due to running tracer '%s'\n",
1987 static void tracing_reset_cpu(struct array_buffer
*buf
, int cpu
)
1989 struct trace_buffer
*buffer
= buf
->buffer
;
1994 ring_buffer_record_disable(buffer
);
1996 /* Make sure all commits have finished */
1998 ring_buffer_reset_cpu(buffer
, cpu
);
2000 ring_buffer_record_enable(buffer
);
2003 void tracing_reset_online_cpus(struct array_buffer
*buf
)
2005 struct trace_buffer
*buffer
= buf
->buffer
;
2011 ring_buffer_record_disable(buffer
);
2013 /* Make sure all commits have finished */
2016 buf
->time_start
= buffer_ftrace_now(buf
, buf
->cpu
);
2018 for_each_online_cpu(cpu
)
2019 ring_buffer_reset_cpu(buffer
, cpu
);
2021 ring_buffer_record_enable(buffer
);
2024 /* Must have trace_types_lock held */
2025 void tracing_reset_all_online_cpus(void)
2027 struct trace_array
*tr
;
2029 list_for_each_entry(tr
, &ftrace_trace_arrays
, list
) {
2030 if (!tr
->clear_trace
)
2032 tr
->clear_trace
= false;
2033 tracing_reset_online_cpus(&tr
->array_buffer
);
2034 #ifdef CONFIG_TRACER_MAX_TRACE
2035 tracing_reset_online_cpus(&tr
->max_buffer
);
2040 static int *tgid_map
;
2042 #define SAVED_CMDLINES_DEFAULT 128
2043 #define NO_CMDLINE_MAP UINT_MAX
2044 static arch_spinlock_t trace_cmdline_lock
= __ARCH_SPIN_LOCK_UNLOCKED
;
2045 struct saved_cmdlines_buffer
{
2046 unsigned map_pid_to_cmdline
[PID_MAX_DEFAULT
+1];
2047 unsigned *map_cmdline_to_pid
;
2048 unsigned cmdline_num
;
2050 char *saved_cmdlines
;
2052 static struct saved_cmdlines_buffer
*savedcmd
;
2054 /* temporary disable recording */
2055 static atomic_t trace_record_taskinfo_disabled __read_mostly
;
2057 static inline char *get_saved_cmdlines(int idx
)
2059 return &savedcmd
->saved_cmdlines
[idx
* TASK_COMM_LEN
];
2062 static inline void set_cmdline(int idx
, const char *cmdline
)
2064 strncpy(get_saved_cmdlines(idx
), cmdline
, TASK_COMM_LEN
);
2067 static int allocate_cmdlines_buffer(unsigned int val
,
2068 struct saved_cmdlines_buffer
*s
)
2070 s
->map_cmdline_to_pid
= kmalloc_array(val
,
2071 sizeof(*s
->map_cmdline_to_pid
),
2073 if (!s
->map_cmdline_to_pid
)
2076 s
->saved_cmdlines
= kmalloc_array(TASK_COMM_LEN
, val
, GFP_KERNEL
);
2077 if (!s
->saved_cmdlines
) {
2078 kfree(s
->map_cmdline_to_pid
);
2083 s
->cmdline_num
= val
;
2084 memset(&s
->map_pid_to_cmdline
, NO_CMDLINE_MAP
,
2085 sizeof(s
->map_pid_to_cmdline
));
2086 memset(s
->map_cmdline_to_pid
, NO_CMDLINE_MAP
,
2087 val
* sizeof(*s
->map_cmdline_to_pid
));
2092 static int trace_create_savedcmd(void)
2096 savedcmd
= kmalloc(sizeof(*savedcmd
), GFP_KERNEL
);
2100 ret
= allocate_cmdlines_buffer(SAVED_CMDLINES_DEFAULT
, savedcmd
);
2110 int is_tracing_stopped(void)
2112 return global_trace
.stop_count
;
2116 * tracing_start - quick start of the tracer
2118 * If tracing is enabled but was stopped by tracing_stop,
2119 * this will start the tracer back up.
2121 void tracing_start(void)
2123 struct trace_buffer
*buffer
;
2124 unsigned long flags
;
2126 if (tracing_disabled
)
2129 raw_spin_lock_irqsave(&global_trace
.start_lock
, flags
);
2130 if (--global_trace
.stop_count
) {
2131 if (global_trace
.stop_count
< 0) {
2132 /* Someone screwed up their debugging */
2134 global_trace
.stop_count
= 0;
2139 /* Prevent the buffers from switching */
2140 arch_spin_lock(&global_trace
.max_lock
);
2142 buffer
= global_trace
.array_buffer
.buffer
;
2144 ring_buffer_record_enable(buffer
);
2146 #ifdef CONFIG_TRACER_MAX_TRACE
2147 buffer
= global_trace
.max_buffer
.buffer
;
2149 ring_buffer_record_enable(buffer
);
2152 arch_spin_unlock(&global_trace
.max_lock
);
2155 raw_spin_unlock_irqrestore(&global_trace
.start_lock
, flags
);
2158 static void tracing_start_tr(struct trace_array
*tr
)
2160 struct trace_buffer
*buffer
;
2161 unsigned long flags
;
2163 if (tracing_disabled
)
2166 /* If global, we need to also start the max tracer */
2167 if (tr
->flags
& TRACE_ARRAY_FL_GLOBAL
)
2168 return tracing_start();
2170 raw_spin_lock_irqsave(&tr
->start_lock
, flags
);
2172 if (--tr
->stop_count
) {
2173 if (tr
->stop_count
< 0) {
2174 /* Someone screwed up their debugging */
2181 buffer
= tr
->array_buffer
.buffer
;
2183 ring_buffer_record_enable(buffer
);
2186 raw_spin_unlock_irqrestore(&tr
->start_lock
, flags
);
2190 * tracing_stop - quick stop of the tracer
2192 * Light weight way to stop tracing. Use in conjunction with
2195 void tracing_stop(void)
2197 struct trace_buffer
*buffer
;
2198 unsigned long flags
;
2200 raw_spin_lock_irqsave(&global_trace
.start_lock
, flags
);
2201 if (global_trace
.stop_count
++)
2204 /* Prevent the buffers from switching */
2205 arch_spin_lock(&global_trace
.max_lock
);
2207 buffer
= global_trace
.array_buffer
.buffer
;
2209 ring_buffer_record_disable(buffer
);
2211 #ifdef CONFIG_TRACER_MAX_TRACE
2212 buffer
= global_trace
.max_buffer
.buffer
;
2214 ring_buffer_record_disable(buffer
);
2217 arch_spin_unlock(&global_trace
.max_lock
);
2220 raw_spin_unlock_irqrestore(&global_trace
.start_lock
, flags
);
2223 static void tracing_stop_tr(struct trace_array
*tr
)
2225 struct trace_buffer
*buffer
;
2226 unsigned long flags
;
2228 /* If global, we need to also stop the max tracer */
2229 if (tr
->flags
& TRACE_ARRAY_FL_GLOBAL
)
2230 return tracing_stop();
2232 raw_spin_lock_irqsave(&tr
->start_lock
, flags
);
2233 if (tr
->stop_count
++)
2236 buffer
= tr
->array_buffer
.buffer
;
2238 ring_buffer_record_disable(buffer
);
2241 raw_spin_unlock_irqrestore(&tr
->start_lock
, flags
);
2244 static int trace_save_cmdline(struct task_struct
*tsk
)
2248 /* treat recording of idle task as a success */
2252 if (unlikely(tsk
->pid
> PID_MAX_DEFAULT
))
2256 * It's not the end of the world if we don't get
2257 * the lock, but we also don't want to spin
2258 * nor do we want to disable interrupts,
2259 * so if we miss here, then better luck next time.
2261 if (!arch_spin_trylock(&trace_cmdline_lock
))
2264 idx
= savedcmd
->map_pid_to_cmdline
[tsk
->pid
];
2265 if (idx
== NO_CMDLINE_MAP
) {
2266 idx
= (savedcmd
->cmdline_idx
+ 1) % savedcmd
->cmdline_num
;
2269 * Check whether the cmdline buffer at idx has a pid
2270 * mapped. We are going to overwrite that entry so we
2271 * need to clear the map_pid_to_cmdline. Otherwise we
2272 * would read the new comm for the old pid.
2274 pid
= savedcmd
->map_cmdline_to_pid
[idx
];
2275 if (pid
!= NO_CMDLINE_MAP
)
2276 savedcmd
->map_pid_to_cmdline
[pid
] = NO_CMDLINE_MAP
;
2278 savedcmd
->map_cmdline_to_pid
[idx
] = tsk
->pid
;
2279 savedcmd
->map_pid_to_cmdline
[tsk
->pid
] = idx
;
2281 savedcmd
->cmdline_idx
= idx
;
2284 set_cmdline(idx
, tsk
->comm
);
2286 arch_spin_unlock(&trace_cmdline_lock
);
2291 static void __trace_find_cmdline(int pid
, char comm
[])
2296 strcpy(comm
, "<idle>");
2300 if (WARN_ON_ONCE(pid
< 0)) {
2301 strcpy(comm
, "<XXX>");
2305 if (pid
> PID_MAX_DEFAULT
) {
2306 strcpy(comm
, "<...>");
2310 map
= savedcmd
->map_pid_to_cmdline
[pid
];
2311 if (map
!= NO_CMDLINE_MAP
)
2312 strlcpy(comm
, get_saved_cmdlines(map
), TASK_COMM_LEN
);
2314 strcpy(comm
, "<...>");
2317 void trace_find_cmdline(int pid
, char comm
[])
2320 arch_spin_lock(&trace_cmdline_lock
);
2322 __trace_find_cmdline(pid
, comm
);
2324 arch_spin_unlock(&trace_cmdline_lock
);
2328 int trace_find_tgid(int pid
)
2330 if (unlikely(!tgid_map
|| !pid
|| pid
> PID_MAX_DEFAULT
))
2333 return tgid_map
[pid
];
2336 static int trace_save_tgid(struct task_struct
*tsk
)
2338 /* treat recording of idle task as a success */
2342 if (unlikely(!tgid_map
|| tsk
->pid
> PID_MAX_DEFAULT
))
2345 tgid_map
[tsk
->pid
] = tsk
->tgid
;
2349 static bool tracing_record_taskinfo_skip(int flags
)
2351 if (unlikely(!(flags
& (TRACE_RECORD_CMDLINE
| TRACE_RECORD_TGID
))))
2353 if (atomic_read(&trace_record_taskinfo_disabled
) || !tracing_is_on())
2355 if (!__this_cpu_read(trace_taskinfo_save
))
2361 * tracing_record_taskinfo - record the task info of a task
2363 * @task: task to record
2364 * @flags: TRACE_RECORD_CMDLINE for recording comm
2365 * TRACE_RECORD_TGID for recording tgid
2367 void tracing_record_taskinfo(struct task_struct
*task
, int flags
)
2371 if (tracing_record_taskinfo_skip(flags
))
2375 * Record as much task information as possible. If some fail, continue
2376 * to try to record the others.
2378 done
= !(flags
& TRACE_RECORD_CMDLINE
) || trace_save_cmdline(task
);
2379 done
&= !(flags
& TRACE_RECORD_TGID
) || trace_save_tgid(task
);
2381 /* If recording any information failed, retry again soon. */
2385 __this_cpu_write(trace_taskinfo_save
, false);
2389 * tracing_record_taskinfo_sched_switch - record task info for sched_switch
2391 * @prev: previous task during sched_switch
2392 * @next: next task during sched_switch
2393 * @flags: TRACE_RECORD_CMDLINE for recording comm
2394 * TRACE_RECORD_TGID for recording tgid
2396 void tracing_record_taskinfo_sched_switch(struct task_struct
*prev
,
2397 struct task_struct
*next
, int flags
)
2401 if (tracing_record_taskinfo_skip(flags
))
2405 * Record as much task information as possible. If some fail, continue
2406 * to try to record the others.
2408 done
= !(flags
& TRACE_RECORD_CMDLINE
) || trace_save_cmdline(prev
);
2409 done
&= !(flags
& TRACE_RECORD_CMDLINE
) || trace_save_cmdline(next
);
2410 done
&= !(flags
& TRACE_RECORD_TGID
) || trace_save_tgid(prev
);
2411 done
&= !(flags
& TRACE_RECORD_TGID
) || trace_save_tgid(next
);
2413 /* If recording any information failed, retry again soon. */
2417 __this_cpu_write(trace_taskinfo_save
, false);
2420 /* Helpers to record a specific task information */
2421 void tracing_record_cmdline(struct task_struct
*task
)
2423 tracing_record_taskinfo(task
, TRACE_RECORD_CMDLINE
);
2426 void tracing_record_tgid(struct task_struct
*task
)
2428 tracing_record_taskinfo(task
, TRACE_RECORD_TGID
);
2432 * Several functions return TRACE_TYPE_PARTIAL_LINE if the trace_seq
2433 * overflowed, and TRACE_TYPE_HANDLED otherwise. This helper function
2434 * simplifies those functions and keeps them in sync.
2436 enum print_line_t
trace_handle_return(struct trace_seq
*s
)
2438 return trace_seq_has_overflowed(s
) ?
2439 TRACE_TYPE_PARTIAL_LINE
: TRACE_TYPE_HANDLED
;
2441 EXPORT_SYMBOL_GPL(trace_handle_return
);
2444 tracing_generic_entry_update(struct trace_entry
*entry
, unsigned short type
,
2445 unsigned long flags
, int pc
)
2447 struct task_struct
*tsk
= current
;
2449 entry
->preempt_count
= pc
& 0xff;
2450 entry
->pid
= (tsk
) ? tsk
->pid
: 0;
2453 #ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT
2454 (irqs_disabled_flags(flags
) ? TRACE_FLAG_IRQS_OFF
: 0) |
2456 TRACE_FLAG_IRQS_NOSUPPORT
|
2458 ((pc
& NMI_MASK
) ? TRACE_FLAG_NMI
: 0) |
2459 ((pc
& HARDIRQ_MASK
) ? TRACE_FLAG_HARDIRQ
: 0) |
2460 ((pc
& SOFTIRQ_OFFSET
) ? TRACE_FLAG_SOFTIRQ
: 0) |
2461 (tif_need_resched() ? TRACE_FLAG_NEED_RESCHED
: 0) |
2462 (test_preempt_need_resched() ? TRACE_FLAG_PREEMPT_RESCHED
: 0);
2464 EXPORT_SYMBOL_GPL(tracing_generic_entry_update
);
2466 struct ring_buffer_event
*
2467 trace_buffer_lock_reserve(struct trace_buffer
*buffer
,
2470 unsigned long flags
, int pc
)
2472 return __trace_buffer_lock_reserve(buffer
, type
, len
, flags
, pc
);
2475 DEFINE_PER_CPU(struct ring_buffer_event
*, trace_buffered_event
);
2476 DEFINE_PER_CPU(int, trace_buffered_event_cnt
);
2477 static int trace_buffered_event_ref
;
2480 * trace_buffered_event_enable - enable buffering events
2482 * When events are being filtered, it is quicker to use a temporary
2483 * buffer to write the event data into if there's a likely chance
2484 * that it will not be committed. The discard of the ring buffer
2485 * is not as fast as committing, and is much slower than copying
2488 * When an event is to be filtered, allocate per cpu buffers to
2489 * write the event data into, and if the event is filtered and discarded
2490 * it is simply dropped, otherwise, the entire data is to be committed
2493 void trace_buffered_event_enable(void)
2495 struct ring_buffer_event
*event
;
2499 WARN_ON_ONCE(!mutex_is_locked(&event_mutex
));
2501 if (trace_buffered_event_ref
++)
2504 for_each_tracing_cpu(cpu
) {
2505 page
= alloc_pages_node(cpu_to_node(cpu
),
2506 GFP_KERNEL
| __GFP_NORETRY
, 0);
2510 event
= page_address(page
);
2511 memset(event
, 0, sizeof(*event
));
2513 per_cpu(trace_buffered_event
, cpu
) = event
;
2516 if (cpu
== smp_processor_id() &&
2517 this_cpu_read(trace_buffered_event
) !=
2518 per_cpu(trace_buffered_event
, cpu
))
2525 trace_buffered_event_disable();
2528 static void enable_trace_buffered_event(void *data
)
2530 /* Probably not needed, but do it anyway */
2532 this_cpu_dec(trace_buffered_event_cnt
);
2535 static void disable_trace_buffered_event(void *data
)
2537 this_cpu_inc(trace_buffered_event_cnt
);
2541 * trace_buffered_event_disable - disable buffering events
2543 * When a filter is removed, it is faster to not use the buffered
2544 * events, and to commit directly into the ring buffer. Free up
2545 * the temp buffers when there are no more users. This requires
2546 * special synchronization with current events.
2548 void trace_buffered_event_disable(void)
2552 WARN_ON_ONCE(!mutex_is_locked(&event_mutex
));
2554 if (WARN_ON_ONCE(!trace_buffered_event_ref
))
2557 if (--trace_buffered_event_ref
)
2561 /* For each CPU, set the buffer as used. */
2562 smp_call_function_many(tracing_buffer_mask
,
2563 disable_trace_buffered_event
, NULL
, 1);
2566 /* Wait for all current users to finish */
2569 for_each_tracing_cpu(cpu
) {
2570 free_page((unsigned long)per_cpu(trace_buffered_event
, cpu
));
2571 per_cpu(trace_buffered_event
, cpu
) = NULL
;
2574 * Make sure trace_buffered_event is NULL before clearing
2575 * trace_buffered_event_cnt.
2580 /* Do the work on each cpu */
2581 smp_call_function_many(tracing_buffer_mask
,
2582 enable_trace_buffered_event
, NULL
, 1);
2586 static struct trace_buffer
*temp_buffer
;
2588 struct ring_buffer_event
*
2589 trace_event_buffer_lock_reserve(struct trace_buffer
**current_rb
,
2590 struct trace_event_file
*trace_file
,
2591 int type
, unsigned long len
,
2592 unsigned long flags
, int pc
)
2594 struct ring_buffer_event
*entry
;
2597 *current_rb
= trace_file
->tr
->array_buffer
.buffer
;
2599 if (!ring_buffer_time_stamp_abs(*current_rb
) && (trace_file
->flags
&
2600 (EVENT_FILE_FL_SOFT_DISABLED
| EVENT_FILE_FL_FILTERED
)) &&
2601 (entry
= this_cpu_read(trace_buffered_event
))) {
2602 /* Try to use the per cpu buffer first */
2603 val
= this_cpu_inc_return(trace_buffered_event_cnt
);
2605 trace_event_setup(entry
, type
, flags
, pc
);
2606 entry
->array
[0] = len
;
2609 this_cpu_dec(trace_buffered_event_cnt
);
2612 entry
= __trace_buffer_lock_reserve(*current_rb
,
2613 type
, len
, flags
, pc
);
2615 * If tracing is off, but we have triggers enabled
2616 * we still need to look at the event data. Use the temp_buffer
2617 * to store the trace event for the tigger to use. It's recusive
2618 * safe and will not be recorded anywhere.
2620 if (!entry
&& trace_file
->flags
& EVENT_FILE_FL_TRIGGER_COND
) {
2621 *current_rb
= temp_buffer
;
2622 entry
= __trace_buffer_lock_reserve(*current_rb
,
2623 type
, len
, flags
, pc
);
2627 EXPORT_SYMBOL_GPL(trace_event_buffer_lock_reserve
);
2629 static DEFINE_SPINLOCK(tracepoint_iter_lock
);
2630 static DEFINE_MUTEX(tracepoint_printk_mutex
);
2632 static void output_printk(struct trace_event_buffer
*fbuffer
)
2634 struct trace_event_call
*event_call
;
2635 struct trace_event_file
*file
;
2636 struct trace_event
*event
;
2637 unsigned long flags
;
2638 struct trace_iterator
*iter
= tracepoint_print_iter
;
2640 /* We should never get here if iter is NULL */
2641 if (WARN_ON_ONCE(!iter
))
2644 event_call
= fbuffer
->trace_file
->event_call
;
2645 if (!event_call
|| !event_call
->event
.funcs
||
2646 !event_call
->event
.funcs
->trace
)
2649 file
= fbuffer
->trace_file
;
2650 if (test_bit(EVENT_FILE_FL_SOFT_DISABLED_BIT
, &file
->flags
) ||
2651 (unlikely(file
->flags
& EVENT_FILE_FL_FILTERED
) &&
2652 !filter_match_preds(file
->filter
, fbuffer
->entry
)))
2655 event
= &fbuffer
->trace_file
->event_call
->event
;
2657 spin_lock_irqsave(&tracepoint_iter_lock
, flags
);
2658 trace_seq_init(&iter
->seq
);
2659 iter
->ent
= fbuffer
->entry
;
2660 event_call
->event
.funcs
->trace(iter
, 0, event
);
2661 trace_seq_putc(&iter
->seq
, 0);
2662 printk("%s", iter
->seq
.buffer
);
2664 spin_unlock_irqrestore(&tracepoint_iter_lock
, flags
);
2667 int tracepoint_printk_sysctl(struct ctl_table
*table
, int write
,
2668 void *buffer
, size_t *lenp
,
2671 int save_tracepoint_printk
;
2674 mutex_lock(&tracepoint_printk_mutex
);
2675 save_tracepoint_printk
= tracepoint_printk
;
2677 ret
= proc_dointvec(table
, write
, buffer
, lenp
, ppos
);
2680 * This will force exiting early, as tracepoint_printk
2681 * is always zero when tracepoint_printk_iter is not allocated
2683 if (!tracepoint_print_iter
)
2684 tracepoint_printk
= 0;
2686 if (save_tracepoint_printk
== tracepoint_printk
)
2689 if (tracepoint_printk
)
2690 static_key_enable(&tracepoint_printk_key
.key
);
2692 static_key_disable(&tracepoint_printk_key
.key
);
2695 mutex_unlock(&tracepoint_printk_mutex
);
2700 void trace_event_buffer_commit(struct trace_event_buffer
*fbuffer
)
2702 if (static_key_false(&tracepoint_printk_key
.key
))
2703 output_printk(fbuffer
);
2705 event_trigger_unlock_commit_regs(fbuffer
->trace_file
, fbuffer
->buffer
,
2706 fbuffer
->event
, fbuffer
->entry
,
2707 fbuffer
->flags
, fbuffer
->pc
, fbuffer
->regs
);
2709 EXPORT_SYMBOL_GPL(trace_event_buffer_commit
);
2714 * trace_buffer_unlock_commit_regs()
2715 * trace_event_buffer_commit()
2716 * trace_event_raw_event_xxx()
2718 # define STACK_SKIP 3
2720 void trace_buffer_unlock_commit_regs(struct trace_array
*tr
,
2721 struct trace_buffer
*buffer
,
2722 struct ring_buffer_event
*event
,
2723 unsigned long flags
, int pc
,
2724 struct pt_regs
*regs
)
2726 __buffer_unlock_commit(buffer
, event
);
2729 * If regs is not set, then skip the necessary functions.
2730 * Note, we can still get here via blktrace, wakeup tracer
2731 * and mmiotrace, but that's ok if they lose a function or
2732 * two. They are not that meaningful.
2734 ftrace_trace_stack(tr
, buffer
, flags
, regs
? 0 : STACK_SKIP
, pc
, regs
);
2735 ftrace_trace_userstack(buffer
, flags
, pc
);
2739 * Similar to trace_buffer_unlock_commit_regs() but do not dump stack.
2742 trace_buffer_unlock_commit_nostack(struct trace_buffer
*buffer
,
2743 struct ring_buffer_event
*event
)
2745 __buffer_unlock_commit(buffer
, event
);
2749 trace_process_export(struct trace_export
*export
,
2750 struct ring_buffer_event
*event
)
2752 struct trace_entry
*entry
;
2753 unsigned int size
= 0;
2755 entry
= ring_buffer_event_data(event
);
2756 size
= ring_buffer_event_length(event
);
2757 export
->write(export
, entry
, size
);
2760 static DEFINE_MUTEX(ftrace_export_lock
);
2762 static struct trace_export __rcu
*ftrace_exports_list __read_mostly
;
2764 static DEFINE_STATIC_KEY_FALSE(ftrace_exports_enabled
);
2766 static inline void ftrace_exports_enable(void)
2768 static_branch_enable(&ftrace_exports_enabled
);
2771 static inline void ftrace_exports_disable(void)
2773 static_branch_disable(&ftrace_exports_enabled
);
2776 static void ftrace_exports(struct ring_buffer_event
*event
)
2778 struct trace_export
*export
;
2780 preempt_disable_notrace();
2782 export
= rcu_dereference_raw_check(ftrace_exports_list
);
2784 trace_process_export(export
, event
);
2785 export
= rcu_dereference_raw_check(export
->next
);
2788 preempt_enable_notrace();
2792 add_trace_export(struct trace_export
**list
, struct trace_export
*export
)
2794 rcu_assign_pointer(export
->next
, *list
);
2796 * We are entering export into the list but another
2797 * CPU might be walking that list. We need to make sure
2798 * the export->next pointer is valid before another CPU sees
2799 * the export pointer included into the list.
2801 rcu_assign_pointer(*list
, export
);
2805 rm_trace_export(struct trace_export
**list
, struct trace_export
*export
)
2807 struct trace_export
**p
;
2809 for (p
= list
; *p
!= NULL
; p
= &(*p
)->next
)
2816 rcu_assign_pointer(*p
, (*p
)->next
);
2822 add_ftrace_export(struct trace_export
**list
, struct trace_export
*export
)
2825 ftrace_exports_enable();
2827 add_trace_export(list
, export
);
2831 rm_ftrace_export(struct trace_export
**list
, struct trace_export
*export
)
2835 ret
= rm_trace_export(list
, export
);
2837 ftrace_exports_disable();
2842 int register_ftrace_export(struct trace_export
*export
)
2844 if (WARN_ON_ONCE(!export
->write
))
2847 mutex_lock(&ftrace_export_lock
);
2849 add_ftrace_export(&ftrace_exports_list
, export
);
2851 mutex_unlock(&ftrace_export_lock
);
2855 EXPORT_SYMBOL_GPL(register_ftrace_export
);
2857 int unregister_ftrace_export(struct trace_export
*export
)
2861 mutex_lock(&ftrace_export_lock
);
2863 ret
= rm_ftrace_export(&ftrace_exports_list
, export
);
2865 mutex_unlock(&ftrace_export_lock
);
2869 EXPORT_SYMBOL_GPL(unregister_ftrace_export
);
2872 trace_function(struct trace_array
*tr
,
2873 unsigned long ip
, unsigned long parent_ip
, unsigned long flags
,
2876 struct trace_event_call
*call
= &event_function
;
2877 struct trace_buffer
*buffer
= tr
->array_buffer
.buffer
;
2878 struct ring_buffer_event
*event
;
2879 struct ftrace_entry
*entry
;
2881 event
= __trace_buffer_lock_reserve(buffer
, TRACE_FN
, sizeof(*entry
),
2885 entry
= ring_buffer_event_data(event
);
2887 entry
->parent_ip
= parent_ip
;
2889 if (!call_filter_check_discard(call
, entry
, buffer
, event
)) {
2890 if (static_branch_unlikely(&ftrace_exports_enabled
))
2891 ftrace_exports(event
);
2892 __buffer_unlock_commit(buffer
, event
);
2896 #ifdef CONFIG_STACKTRACE
2898 /* Allow 4 levels of nesting: normal, softirq, irq, NMI */
2899 #define FTRACE_KSTACK_NESTING 4
2901 #define FTRACE_KSTACK_ENTRIES (PAGE_SIZE / FTRACE_KSTACK_NESTING)
2903 struct ftrace_stack
{
2904 unsigned long calls
[FTRACE_KSTACK_ENTRIES
];
2908 struct ftrace_stacks
{
2909 struct ftrace_stack stacks
[FTRACE_KSTACK_NESTING
];
2912 static DEFINE_PER_CPU(struct ftrace_stacks
, ftrace_stacks
);
2913 static DEFINE_PER_CPU(int, ftrace_stack_reserve
);
2915 static void __ftrace_trace_stack(struct trace_buffer
*buffer
,
2916 unsigned long flags
,
2917 int skip
, int pc
, struct pt_regs
*regs
)
2919 struct trace_event_call
*call
= &event_kernel_stack
;
2920 struct ring_buffer_event
*event
;
2921 unsigned int size
, nr_entries
;
2922 struct ftrace_stack
*fstack
;
2923 struct stack_entry
*entry
;
2927 * Add one, for this function and the call to save_stack_trace()
2928 * If regs is set, then these functions will not be in the way.
2930 #ifndef CONFIG_UNWINDER_ORC
2936 * Since events can happen in NMIs there's no safe way to
2937 * use the per cpu ftrace_stacks. We reserve it and if an interrupt
2938 * or NMI comes in, it will just have to use the default
2939 * FTRACE_STACK_SIZE.
2941 preempt_disable_notrace();
2943 stackidx
= __this_cpu_inc_return(ftrace_stack_reserve
) - 1;
2945 /* This should never happen. If it does, yell once and skip */
2946 if (WARN_ON_ONCE(stackidx
> FTRACE_KSTACK_NESTING
))
2950 * The above __this_cpu_inc_return() is 'atomic' cpu local. An
2951 * interrupt will either see the value pre increment or post
2952 * increment. If the interrupt happens pre increment it will have
2953 * restored the counter when it returns. We just need a barrier to
2954 * keep gcc from moving things around.
2958 fstack
= this_cpu_ptr(ftrace_stacks
.stacks
) + stackidx
;
2959 size
= ARRAY_SIZE(fstack
->calls
);
2962 nr_entries
= stack_trace_save_regs(regs
, fstack
->calls
,
2965 nr_entries
= stack_trace_save(fstack
->calls
, size
, skip
);
2968 size
= nr_entries
* sizeof(unsigned long);
2969 event
= __trace_buffer_lock_reserve(buffer
, TRACE_STACK
,
2970 sizeof(*entry
) + size
, flags
, pc
);
2973 entry
= ring_buffer_event_data(event
);
2975 memcpy(&entry
->caller
, fstack
->calls
, size
);
2976 entry
->size
= nr_entries
;
2978 if (!call_filter_check_discard(call
, entry
, buffer
, event
))
2979 __buffer_unlock_commit(buffer
, event
);
2982 /* Again, don't let gcc optimize things here */
2984 __this_cpu_dec(ftrace_stack_reserve
);
2985 preempt_enable_notrace();
2989 static inline void ftrace_trace_stack(struct trace_array
*tr
,
2990 struct trace_buffer
*buffer
,
2991 unsigned long flags
,
2992 int skip
, int pc
, struct pt_regs
*regs
)
2994 if (!(tr
->trace_flags
& TRACE_ITER_STACKTRACE
))
2997 __ftrace_trace_stack(buffer
, flags
, skip
, pc
, regs
);
3000 void __trace_stack(struct trace_array
*tr
, unsigned long flags
, int skip
,
3003 struct trace_buffer
*buffer
= tr
->array_buffer
.buffer
;
3005 if (rcu_is_watching()) {
3006 __ftrace_trace_stack(buffer
, flags
, skip
, pc
, NULL
);
3011 * When an NMI triggers, RCU is enabled via rcu_nmi_enter(),
3012 * but if the above rcu_is_watching() failed, then the NMI
3013 * triggered someplace critical, and rcu_irq_enter() should
3014 * not be called from NMI.
3016 if (unlikely(in_nmi()))
3019 rcu_irq_enter_irqson();
3020 __ftrace_trace_stack(buffer
, flags
, skip
, pc
, NULL
);
3021 rcu_irq_exit_irqson();
3025 * trace_dump_stack - record a stack back trace in the trace buffer
3026 * @skip: Number of functions to skip (helper handlers)
3028 void trace_dump_stack(int skip
)
3030 unsigned long flags
;
3032 if (tracing_disabled
|| tracing_selftest_running
)
3035 local_save_flags(flags
);
3037 #ifndef CONFIG_UNWINDER_ORC
3038 /* Skip 1 to skip this function. */
3041 __ftrace_trace_stack(global_trace
.array_buffer
.buffer
,
3042 flags
, skip
, preempt_count(), NULL
);
3044 EXPORT_SYMBOL_GPL(trace_dump_stack
);
3046 #ifdef CONFIG_USER_STACKTRACE_SUPPORT
3047 static DEFINE_PER_CPU(int, user_stack_count
);
3050 ftrace_trace_userstack(struct trace_buffer
*buffer
, unsigned long flags
, int pc
)
3052 struct trace_event_call
*call
= &event_user_stack
;
3053 struct ring_buffer_event
*event
;
3054 struct userstack_entry
*entry
;
3056 if (!(global_trace
.trace_flags
& TRACE_ITER_USERSTACKTRACE
))
3060 * NMIs can not handle page faults, even with fix ups.
3061 * The save user stack can (and often does) fault.
3063 if (unlikely(in_nmi()))
3067 * prevent recursion, since the user stack tracing may
3068 * trigger other kernel events.
3071 if (__this_cpu_read(user_stack_count
))
3074 __this_cpu_inc(user_stack_count
);
3076 event
= __trace_buffer_lock_reserve(buffer
, TRACE_USER_STACK
,
3077 sizeof(*entry
), flags
, pc
);
3079 goto out_drop_count
;
3080 entry
= ring_buffer_event_data(event
);
3082 entry
->tgid
= current
->tgid
;
3083 memset(&entry
->caller
, 0, sizeof(entry
->caller
));
3085 stack_trace_save_user(entry
->caller
, FTRACE_STACK_ENTRIES
);
3086 if (!call_filter_check_discard(call
, entry
, buffer
, event
))
3087 __buffer_unlock_commit(buffer
, event
);
3090 __this_cpu_dec(user_stack_count
);
3094 #else /* CONFIG_USER_STACKTRACE_SUPPORT */
3095 static void ftrace_trace_userstack(struct trace_buffer
*buffer
,
3096 unsigned long flags
, int pc
)
3099 #endif /* !CONFIG_USER_STACKTRACE_SUPPORT */
3101 #endif /* CONFIG_STACKTRACE */
3103 /* created for use with alloc_percpu */
3104 struct trace_buffer_struct
{
3106 char buffer
[4][TRACE_BUF_SIZE
];
3109 static struct trace_buffer_struct
*trace_percpu_buffer
;
3112 * Thise allows for lockless recording. If we're nested too deeply, then
3113 * this returns NULL.
3115 static char *get_trace_buf(void)
3117 struct trace_buffer_struct
*buffer
= this_cpu_ptr(trace_percpu_buffer
);
3119 if (!buffer
|| buffer
->nesting
>= 4)
3124 /* Interrupts must see nesting incremented before we use the buffer */
3126 return &buffer
->buffer
[buffer
->nesting
][0];
3129 static void put_trace_buf(void)
3131 /* Don't let the decrement of nesting leak before this */
3133 this_cpu_dec(trace_percpu_buffer
->nesting
);
3136 static int alloc_percpu_trace_buffer(void)
3138 struct trace_buffer_struct
*buffers
;
3140 buffers
= alloc_percpu(struct trace_buffer_struct
);
3141 if (MEM_FAIL(!buffers
, "Could not allocate percpu trace_printk buffer"))
3144 trace_percpu_buffer
= buffers
;
3148 static int buffers_allocated
;
3150 void trace_printk_init_buffers(void)
3152 if (buffers_allocated
)
3155 if (alloc_percpu_trace_buffer())
3158 /* trace_printk() is for debug use only. Don't use it in production. */
3161 pr_warn("**********************************************************\n");
3162 pr_warn("** NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE **\n");
3164 pr_warn("** trace_printk() being used. Allocating extra memory. **\n");
3166 pr_warn("** This means that this is a DEBUG kernel and it is **\n");
3167 pr_warn("** unsafe for production use. **\n");
3169 pr_warn("** If you see this message and you are not debugging **\n");
3170 pr_warn("** the kernel, report this immediately to your vendor! **\n");
3172 pr_warn("** NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE **\n");
3173 pr_warn("**********************************************************\n");
3175 /* Expand the buffers to set size */
3176 tracing_update_buffers();
3178 buffers_allocated
= 1;
3181 * trace_printk_init_buffers() can be called by modules.
3182 * If that happens, then we need to start cmdline recording
3183 * directly here. If the global_trace.buffer is already
3184 * allocated here, then this was called by module code.
3186 if (global_trace
.array_buffer
.buffer
)
3187 tracing_start_cmdline_record();
3189 EXPORT_SYMBOL_GPL(trace_printk_init_buffers
);
3191 void trace_printk_start_comm(void)
3193 /* Start tracing comms if trace printk is set */
3194 if (!buffers_allocated
)
3196 tracing_start_cmdline_record();
3199 static void trace_printk_start_stop_comm(int enabled
)
3201 if (!buffers_allocated
)
3205 tracing_start_cmdline_record();
3207 tracing_stop_cmdline_record();
3211 * trace_vbprintk - write binary msg to tracing buffer
3212 * @ip: The address of the caller
3213 * @fmt: The string format to write to the buffer
3214 * @args: Arguments for @fmt
3216 int trace_vbprintk(unsigned long ip
, const char *fmt
, va_list args
)
3218 struct trace_event_call
*call
= &event_bprint
;
3219 struct ring_buffer_event
*event
;
3220 struct trace_buffer
*buffer
;
3221 struct trace_array
*tr
= &global_trace
;
3222 struct bprint_entry
*entry
;
3223 unsigned long flags
;
3225 int len
= 0, size
, pc
;
3227 if (unlikely(tracing_selftest_running
|| tracing_disabled
))
3230 /* Don't pollute graph traces with trace_vprintk internals */
3231 pause_graph_tracing();
3233 pc
= preempt_count();
3234 preempt_disable_notrace();
3236 tbuffer
= get_trace_buf();
3242 len
= vbin_printf((u32
*)tbuffer
, TRACE_BUF_SIZE
/sizeof(int), fmt
, args
);
3244 if (len
> TRACE_BUF_SIZE
/sizeof(int) || len
< 0)
3247 local_save_flags(flags
);
3248 size
= sizeof(*entry
) + sizeof(u32
) * len
;
3249 buffer
= tr
->array_buffer
.buffer
;
3250 ring_buffer_nest_start(buffer
);
3251 event
= __trace_buffer_lock_reserve(buffer
, TRACE_BPRINT
, size
,
3255 entry
= ring_buffer_event_data(event
);
3259 memcpy(entry
->buf
, tbuffer
, sizeof(u32
) * len
);
3260 if (!call_filter_check_discard(call
, entry
, buffer
, event
)) {
3261 __buffer_unlock_commit(buffer
, event
);
3262 ftrace_trace_stack(tr
, buffer
, flags
, 6, pc
, NULL
);
3266 ring_buffer_nest_end(buffer
);
3271 preempt_enable_notrace();
3272 unpause_graph_tracing();
3276 EXPORT_SYMBOL_GPL(trace_vbprintk
);
3280 __trace_array_vprintk(struct trace_buffer
*buffer
,
3281 unsigned long ip
, const char *fmt
, va_list args
)
3283 struct trace_event_call
*call
= &event_print
;
3284 struct ring_buffer_event
*event
;
3285 int len
= 0, size
, pc
;
3286 struct print_entry
*entry
;
3287 unsigned long flags
;
3290 if (tracing_disabled
|| tracing_selftest_running
)
3293 /* Don't pollute graph traces with trace_vprintk internals */
3294 pause_graph_tracing();
3296 pc
= preempt_count();
3297 preempt_disable_notrace();
3300 tbuffer
= get_trace_buf();
3306 len
= vscnprintf(tbuffer
, TRACE_BUF_SIZE
, fmt
, args
);
3308 local_save_flags(flags
);
3309 size
= sizeof(*entry
) + len
+ 1;
3310 ring_buffer_nest_start(buffer
);
3311 event
= __trace_buffer_lock_reserve(buffer
, TRACE_PRINT
, size
,
3315 entry
= ring_buffer_event_data(event
);
3318 memcpy(&entry
->buf
, tbuffer
, len
+ 1);
3319 if (!call_filter_check_discard(call
, entry
, buffer
, event
)) {
3320 __buffer_unlock_commit(buffer
, event
);
3321 ftrace_trace_stack(&global_trace
, buffer
, flags
, 6, pc
, NULL
);
3325 ring_buffer_nest_end(buffer
);
3329 preempt_enable_notrace();
3330 unpause_graph_tracing();
3336 int trace_array_vprintk(struct trace_array
*tr
,
3337 unsigned long ip
, const char *fmt
, va_list args
)
3339 return __trace_array_vprintk(tr
->array_buffer
.buffer
, ip
, fmt
, args
);
3343 int trace_array_printk(struct trace_array
*tr
,
3344 unsigned long ip
, const char *fmt
, ...)
3349 if (!(global_trace
.trace_flags
& TRACE_ITER_PRINTK
))
3356 ret
= trace_array_vprintk(tr
, ip
, fmt
, ap
);
3360 EXPORT_SYMBOL_GPL(trace_array_printk
);
3363 int trace_array_printk_buf(struct trace_buffer
*buffer
,
3364 unsigned long ip
, const char *fmt
, ...)
3369 if (!(global_trace
.trace_flags
& TRACE_ITER_PRINTK
))
3373 ret
= __trace_array_vprintk(buffer
, ip
, fmt
, ap
);
3379 int trace_vprintk(unsigned long ip
, const char *fmt
, va_list args
)
3381 return trace_array_vprintk(&global_trace
, ip
, fmt
, args
);
3383 EXPORT_SYMBOL_GPL(trace_vprintk
);
3385 static void trace_iterator_increment(struct trace_iterator
*iter
)
3387 struct ring_buffer_iter
*buf_iter
= trace_buffer_iter(iter
, iter
->cpu
);
3391 ring_buffer_iter_advance(buf_iter
);
3394 static struct trace_entry
*
3395 peek_next_entry(struct trace_iterator
*iter
, int cpu
, u64
*ts
,
3396 unsigned long *lost_events
)
3398 struct ring_buffer_event
*event
;
3399 struct ring_buffer_iter
*buf_iter
= trace_buffer_iter(iter
, cpu
);
3402 event
= ring_buffer_iter_peek(buf_iter
, ts
);
3404 *lost_events
= ring_buffer_iter_dropped(buf_iter
) ?
3405 (unsigned long)-1 : 0;
3407 event
= ring_buffer_peek(iter
->array_buffer
->buffer
, cpu
, ts
,
3412 iter
->ent_size
= ring_buffer_event_length(event
);
3413 return ring_buffer_event_data(event
);
3419 static struct trace_entry
*
3420 __find_next_entry(struct trace_iterator
*iter
, int *ent_cpu
,
3421 unsigned long *missing_events
, u64
*ent_ts
)
3423 struct trace_buffer
*buffer
= iter
->array_buffer
->buffer
;
3424 struct trace_entry
*ent
, *next
= NULL
;
3425 unsigned long lost_events
= 0, next_lost
= 0;
3426 int cpu_file
= iter
->cpu_file
;
3427 u64 next_ts
= 0, ts
;
3433 * If we are in a per_cpu trace file, don't bother by iterating over
3434 * all cpu and peek directly.
3436 if (cpu_file
> RING_BUFFER_ALL_CPUS
) {
3437 if (ring_buffer_empty_cpu(buffer
, cpu_file
))
3439 ent
= peek_next_entry(iter
, cpu_file
, ent_ts
, missing_events
);
3441 *ent_cpu
= cpu_file
;
3446 for_each_tracing_cpu(cpu
) {
3448 if (ring_buffer_empty_cpu(buffer
, cpu
))
3451 ent
= peek_next_entry(iter
, cpu
, &ts
, &lost_events
);
3454 * Pick the entry with the smallest timestamp:
3456 if (ent
&& (!next
|| ts
< next_ts
)) {
3460 next_lost
= lost_events
;
3461 next_size
= iter
->ent_size
;
3465 iter
->ent_size
= next_size
;
3468 *ent_cpu
= next_cpu
;
3474 *missing_events
= next_lost
;
3479 #define STATIC_TEMP_BUF_SIZE 128
3480 static char static_temp_buf
[STATIC_TEMP_BUF_SIZE
];
3482 /* Find the next real entry, without updating the iterator itself */
3483 struct trace_entry
*trace_find_next_entry(struct trace_iterator
*iter
,
3484 int *ent_cpu
, u64
*ent_ts
)
3486 /* __find_next_entry will reset ent_size */
3487 int ent_size
= iter
->ent_size
;
3488 struct trace_entry
*entry
;
3491 * If called from ftrace_dump(), then the iter->temp buffer
3492 * will be the static_temp_buf and not created from kmalloc.
3493 * If the entry size is greater than the buffer, we can
3494 * not save it. Just return NULL in that case. This is only
3495 * used to add markers when two consecutive events' time
3496 * stamps have a large delta. See trace_print_lat_context()
3498 if (iter
->temp
== static_temp_buf
&&
3499 STATIC_TEMP_BUF_SIZE
< ent_size
)
3503 * The __find_next_entry() may call peek_next_entry(), which may
3504 * call ring_buffer_peek() that may make the contents of iter->ent
3505 * undefined. Need to copy iter->ent now.
3507 if (iter
->ent
&& iter
->ent
!= iter
->temp
) {
3508 if ((!iter
->temp
|| iter
->temp_size
< iter
->ent_size
) &&
3509 !WARN_ON_ONCE(iter
->temp
== static_temp_buf
)) {
3511 iter
->temp
= kmalloc(iter
->ent_size
, GFP_KERNEL
);
3515 memcpy(iter
->temp
, iter
->ent
, iter
->ent_size
);
3516 iter
->temp_size
= iter
->ent_size
;
3517 iter
->ent
= iter
->temp
;
3519 entry
= __find_next_entry(iter
, ent_cpu
, NULL
, ent_ts
);
3520 /* Put back the original ent_size */
3521 iter
->ent_size
= ent_size
;
3526 /* Find the next real entry, and increment the iterator to the next entry */
3527 void *trace_find_next_entry_inc(struct trace_iterator
*iter
)
3529 iter
->ent
= __find_next_entry(iter
, &iter
->cpu
,
3530 &iter
->lost_events
, &iter
->ts
);
3533 trace_iterator_increment(iter
);
3535 return iter
->ent
? iter
: NULL
;
3538 static void trace_consume(struct trace_iterator
*iter
)
3540 ring_buffer_consume(iter
->array_buffer
->buffer
, iter
->cpu
, &iter
->ts
,
3541 &iter
->lost_events
);
3544 static void *s_next(struct seq_file
*m
, void *v
, loff_t
*pos
)
3546 struct trace_iterator
*iter
= m
->private;
3550 WARN_ON_ONCE(iter
->leftover
);
3554 /* can't go backwards */
3559 ent
= trace_find_next_entry_inc(iter
);
3563 while (ent
&& iter
->idx
< i
)
3564 ent
= trace_find_next_entry_inc(iter
);
3571 void tracing_iter_reset(struct trace_iterator
*iter
, int cpu
)
3573 struct ring_buffer_event
*event
;
3574 struct ring_buffer_iter
*buf_iter
;
3575 unsigned long entries
= 0;
3578 per_cpu_ptr(iter
->array_buffer
->data
, cpu
)->skipped_entries
= 0;
3580 buf_iter
= trace_buffer_iter(iter
, cpu
);
3584 ring_buffer_iter_reset(buf_iter
);
3587 * We could have the case with the max latency tracers
3588 * that a reset never took place on a cpu. This is evident
3589 * by the timestamp being before the start of the buffer.
3591 while ((event
= ring_buffer_iter_peek(buf_iter
, &ts
))) {
3592 if (ts
>= iter
->array_buffer
->time_start
)
3595 ring_buffer_iter_advance(buf_iter
);
3598 per_cpu_ptr(iter
->array_buffer
->data
, cpu
)->skipped_entries
= entries
;
3602 * The current tracer is copied to avoid a global locking
3605 static void *s_start(struct seq_file
*m
, loff_t
*pos
)
3607 struct trace_iterator
*iter
= m
->private;
3608 struct trace_array
*tr
= iter
->tr
;
3609 int cpu_file
= iter
->cpu_file
;
3615 * copy the tracer to avoid using a global lock all around.
3616 * iter->trace is a copy of current_trace, the pointer to the
3617 * name may be used instead of a strcmp(), as iter->trace->name
3618 * will point to the same string as current_trace->name.
3620 mutex_lock(&trace_types_lock
);
3621 if (unlikely(tr
->current_trace
&& iter
->trace
->name
!= tr
->current_trace
->name
))
3622 *iter
->trace
= *tr
->current_trace
;
3623 mutex_unlock(&trace_types_lock
);
3625 #ifdef CONFIG_TRACER_MAX_TRACE
3626 if (iter
->snapshot
&& iter
->trace
->use_max_tr
)
3627 return ERR_PTR(-EBUSY
);
3630 if (!iter
->snapshot
)
3631 atomic_inc(&trace_record_taskinfo_disabled
);
3633 if (*pos
!= iter
->pos
) {
3638 if (cpu_file
== RING_BUFFER_ALL_CPUS
) {
3639 for_each_tracing_cpu(cpu
)
3640 tracing_iter_reset(iter
, cpu
);
3642 tracing_iter_reset(iter
, cpu_file
);
3645 for (p
= iter
; p
&& l
< *pos
; p
= s_next(m
, p
, &l
))
3650 * If we overflowed the seq_file before, then we want
3651 * to just reuse the trace_seq buffer again.
3657 p
= s_next(m
, p
, &l
);
3661 trace_event_read_lock();
3662 trace_access_lock(cpu_file
);
3666 static void s_stop(struct seq_file
*m
, void *p
)
3668 struct trace_iterator
*iter
= m
->private;
3670 #ifdef CONFIG_TRACER_MAX_TRACE
3671 if (iter
->snapshot
&& iter
->trace
->use_max_tr
)
3675 if (!iter
->snapshot
)
3676 atomic_dec(&trace_record_taskinfo_disabled
);
3678 trace_access_unlock(iter
->cpu_file
);
3679 trace_event_read_unlock();
3683 get_total_entries_cpu(struct array_buffer
*buf
, unsigned long *total
,
3684 unsigned long *entries
, int cpu
)
3686 unsigned long count
;
3688 count
= ring_buffer_entries_cpu(buf
->buffer
, cpu
);
3690 * If this buffer has skipped entries, then we hold all
3691 * entries for the trace and we need to ignore the
3692 * ones before the time stamp.
3694 if (per_cpu_ptr(buf
->data
, cpu
)->skipped_entries
) {
3695 count
-= per_cpu_ptr(buf
->data
, cpu
)->skipped_entries
;
3696 /* total is the same as the entries */
3700 ring_buffer_overrun_cpu(buf
->buffer
, cpu
);
3705 get_total_entries(struct array_buffer
*buf
,
3706 unsigned long *total
, unsigned long *entries
)
3714 for_each_tracing_cpu(cpu
) {
3715 get_total_entries_cpu(buf
, &t
, &e
, cpu
);
3721 unsigned long trace_total_entries_cpu(struct trace_array
*tr
, int cpu
)
3723 unsigned long total
, entries
;
3728 get_total_entries_cpu(&tr
->array_buffer
, &total
, &entries
, cpu
);
3733 unsigned long trace_total_entries(struct trace_array
*tr
)
3735 unsigned long total
, entries
;
3740 get_total_entries(&tr
->array_buffer
, &total
, &entries
);
3745 static void print_lat_help_header(struct seq_file
*m
)
3747 seq_puts(m
, "# _------=> CPU# \n"
3748 "# / _-----=> irqs-off \n"
3749 "# | / _----=> need-resched \n"
3750 "# || / _---=> hardirq/softirq \n"
3751 "# ||| / _--=> preempt-depth \n"
3753 "# cmd pid ||||| time | caller \n"
3754 "# \\ / ||||| \\ | / \n");
3757 static void print_event_info(struct array_buffer
*buf
, struct seq_file
*m
)
3759 unsigned long total
;
3760 unsigned long entries
;
3762 get_total_entries(buf
, &total
, &entries
);
3763 seq_printf(m
, "# entries-in-buffer/entries-written: %lu/%lu #P:%d\n",
3764 entries
, total
, num_online_cpus());
3768 static void print_func_help_header(struct array_buffer
*buf
, struct seq_file
*m
,
3771 bool tgid
= flags
& TRACE_ITER_RECORD_TGID
;
3773 print_event_info(buf
, m
);
3775 seq_printf(m
, "# TASK-PID %s CPU# TIMESTAMP FUNCTION\n", tgid
? "TGID " : "");
3776 seq_printf(m
, "# | | %s | | |\n", tgid
? " | " : "");
3779 static void print_func_help_header_irq(struct array_buffer
*buf
, struct seq_file
*m
,
3782 bool tgid
= flags
& TRACE_ITER_RECORD_TGID
;
3783 const char *space
= " ";
3784 int prec
= tgid
? 10 : 2;
3786 print_event_info(buf
, m
);
3788 seq_printf(m
, "# %.*s _-----=> irqs-off\n", prec
, space
);
3789 seq_printf(m
, "# %.*s / _----=> need-resched\n", prec
, space
);
3790 seq_printf(m
, "# %.*s| / _---=> hardirq/softirq\n", prec
, space
);
3791 seq_printf(m
, "# %.*s|| / _--=> preempt-depth\n", prec
, space
);
3792 seq_printf(m
, "# %.*s||| / delay\n", prec
, space
);
3793 seq_printf(m
, "# TASK-PID %.*sCPU# |||| TIMESTAMP FUNCTION\n", prec
, " TGID ");
3794 seq_printf(m
, "# | | %.*s | |||| | |\n", prec
, " | ");
3798 print_trace_header(struct seq_file
*m
, struct trace_iterator
*iter
)
3800 unsigned long sym_flags
= (global_trace
.trace_flags
& TRACE_ITER_SYM_MASK
);
3801 struct array_buffer
*buf
= iter
->array_buffer
;
3802 struct trace_array_cpu
*data
= per_cpu_ptr(buf
->data
, buf
->cpu
);
3803 struct tracer
*type
= iter
->trace
;
3804 unsigned long entries
;
3805 unsigned long total
;
3806 const char *name
= "preemption";
3810 get_total_entries(buf
, &total
, &entries
);
3812 seq_printf(m
, "# %s latency trace v1.1.5 on %s\n",
3814 seq_puts(m
, "# -----------------------------------"
3815 "---------------------------------\n");
3816 seq_printf(m
, "# latency: %lu us, #%lu/%lu, CPU#%d |"
3817 " (M:%s VP:%d, KP:%d, SP:%d HP:%d",
3818 nsecs_to_usecs(data
->saved_latency
),
3822 #if defined(CONFIG_PREEMPT_NONE)
3824 #elif defined(CONFIG_PREEMPT_VOLUNTARY)
3826 #elif defined(CONFIG_PREEMPT)
3828 #elif defined(CONFIG_PREEMPT_RT)
3833 /* These are reserved for later use */
3836 seq_printf(m
, " #P:%d)\n", num_online_cpus());
3840 seq_puts(m
, "# -----------------\n");
3841 seq_printf(m
, "# | task: %.16s-%d "
3842 "(uid:%d nice:%ld policy:%ld rt_prio:%ld)\n",
3843 data
->comm
, data
->pid
,
3844 from_kuid_munged(seq_user_ns(m
), data
->uid
), data
->nice
,
3845 data
->policy
, data
->rt_priority
);
3846 seq_puts(m
, "# -----------------\n");
3848 if (data
->critical_start
) {
3849 seq_puts(m
, "# => started at: ");
3850 seq_print_ip_sym(&iter
->seq
, data
->critical_start
, sym_flags
);
3851 trace_print_seq(m
, &iter
->seq
);
3852 seq_puts(m
, "\n# => ended at: ");
3853 seq_print_ip_sym(&iter
->seq
, data
->critical_end
, sym_flags
);
3854 trace_print_seq(m
, &iter
->seq
);
3855 seq_puts(m
, "\n#\n");
3861 static void test_cpu_buff_start(struct trace_iterator
*iter
)
3863 struct trace_seq
*s
= &iter
->seq
;
3864 struct trace_array
*tr
= iter
->tr
;
3866 if (!(tr
->trace_flags
& TRACE_ITER_ANNOTATE
))
3869 if (!(iter
->iter_flags
& TRACE_FILE_ANNOTATE
))
3872 if (cpumask_available(iter
->started
) &&
3873 cpumask_test_cpu(iter
->cpu
, iter
->started
))
3876 if (per_cpu_ptr(iter
->array_buffer
->data
, iter
->cpu
)->skipped_entries
)
3879 if (cpumask_available(iter
->started
))
3880 cpumask_set_cpu(iter
->cpu
, iter
->started
);
3882 /* Don't print started cpu buffer for the first entry of the trace */
3884 trace_seq_printf(s
, "##### CPU %u buffer started ####\n",
3888 static enum print_line_t
print_trace_fmt(struct trace_iterator
*iter
)
3890 struct trace_array
*tr
= iter
->tr
;
3891 struct trace_seq
*s
= &iter
->seq
;
3892 unsigned long sym_flags
= (tr
->trace_flags
& TRACE_ITER_SYM_MASK
);
3893 struct trace_entry
*entry
;
3894 struct trace_event
*event
;
3898 test_cpu_buff_start(iter
);
3900 event
= ftrace_find_event(entry
->type
);
3902 if (tr
->trace_flags
& TRACE_ITER_CONTEXT_INFO
) {
3903 if (iter
->iter_flags
& TRACE_FILE_LAT_FMT
)
3904 trace_print_lat_context(iter
);
3906 trace_print_context(iter
);
3909 if (trace_seq_has_overflowed(s
))
3910 return TRACE_TYPE_PARTIAL_LINE
;
3913 return event
->funcs
->trace(iter
, sym_flags
, event
);
3915 trace_seq_printf(s
, "Unknown type %d\n", entry
->type
);
3917 return trace_handle_return(s
);
3920 static enum print_line_t
print_raw_fmt(struct trace_iterator
*iter
)
3922 struct trace_array
*tr
= iter
->tr
;
3923 struct trace_seq
*s
= &iter
->seq
;
3924 struct trace_entry
*entry
;
3925 struct trace_event
*event
;
3929 if (tr
->trace_flags
& TRACE_ITER_CONTEXT_INFO
)
3930 trace_seq_printf(s
, "%d %d %llu ",
3931 entry
->pid
, iter
->cpu
, iter
->ts
);
3933 if (trace_seq_has_overflowed(s
))
3934 return TRACE_TYPE_PARTIAL_LINE
;
3936 event
= ftrace_find_event(entry
->type
);
3938 return event
->funcs
->raw(iter
, 0, event
);
3940 trace_seq_printf(s
, "%d ?\n", entry
->type
);
3942 return trace_handle_return(s
);
3945 static enum print_line_t
print_hex_fmt(struct trace_iterator
*iter
)
3947 struct trace_array
*tr
= iter
->tr
;
3948 struct trace_seq
*s
= &iter
->seq
;
3949 unsigned char newline
= '\n';
3950 struct trace_entry
*entry
;
3951 struct trace_event
*event
;
3955 if (tr
->trace_flags
& TRACE_ITER_CONTEXT_INFO
) {
3956 SEQ_PUT_HEX_FIELD(s
, entry
->pid
);
3957 SEQ_PUT_HEX_FIELD(s
, iter
->cpu
);
3958 SEQ_PUT_HEX_FIELD(s
, iter
->ts
);
3959 if (trace_seq_has_overflowed(s
))
3960 return TRACE_TYPE_PARTIAL_LINE
;
3963 event
= ftrace_find_event(entry
->type
);
3965 enum print_line_t ret
= event
->funcs
->hex(iter
, 0, event
);
3966 if (ret
!= TRACE_TYPE_HANDLED
)
3970 SEQ_PUT_FIELD(s
, newline
);
3972 return trace_handle_return(s
);
3975 static enum print_line_t
print_bin_fmt(struct trace_iterator
*iter
)
3977 struct trace_array
*tr
= iter
->tr
;
3978 struct trace_seq
*s
= &iter
->seq
;
3979 struct trace_entry
*entry
;
3980 struct trace_event
*event
;
3984 if (tr
->trace_flags
& TRACE_ITER_CONTEXT_INFO
) {
3985 SEQ_PUT_FIELD(s
, entry
->pid
);
3986 SEQ_PUT_FIELD(s
, iter
->cpu
);
3987 SEQ_PUT_FIELD(s
, iter
->ts
);
3988 if (trace_seq_has_overflowed(s
))
3989 return TRACE_TYPE_PARTIAL_LINE
;
3992 event
= ftrace_find_event(entry
->type
);
3993 return event
? event
->funcs
->binary(iter
, 0, event
) :
3997 int trace_empty(struct trace_iterator
*iter
)
3999 struct ring_buffer_iter
*buf_iter
;
4002 /* If we are looking at one CPU buffer, only check that one */
4003 if (iter
->cpu_file
!= RING_BUFFER_ALL_CPUS
) {
4004 cpu
= iter
->cpu_file
;
4005 buf_iter
= trace_buffer_iter(iter
, cpu
);
4007 if (!ring_buffer_iter_empty(buf_iter
))
4010 if (!ring_buffer_empty_cpu(iter
->array_buffer
->buffer
, cpu
))
4016 for_each_tracing_cpu(cpu
) {
4017 buf_iter
= trace_buffer_iter(iter
, cpu
);
4019 if (!ring_buffer_iter_empty(buf_iter
))
4022 if (!ring_buffer_empty_cpu(iter
->array_buffer
->buffer
, cpu
))
4030 /* Called with trace_event_read_lock() held. */
4031 enum print_line_t
print_trace_line(struct trace_iterator
*iter
)
4033 struct trace_array
*tr
= iter
->tr
;
4034 unsigned long trace_flags
= tr
->trace_flags
;
4035 enum print_line_t ret
;
4037 if (iter
->lost_events
) {
4038 if (iter
->lost_events
== (unsigned long)-1)
4039 trace_seq_printf(&iter
->seq
, "CPU:%d [LOST EVENTS]\n",
4042 trace_seq_printf(&iter
->seq
, "CPU:%d [LOST %lu EVENTS]\n",
4043 iter
->cpu
, iter
->lost_events
);
4044 if (trace_seq_has_overflowed(&iter
->seq
))
4045 return TRACE_TYPE_PARTIAL_LINE
;
4048 if (iter
->trace
&& iter
->trace
->print_line
) {
4049 ret
= iter
->trace
->print_line(iter
);
4050 if (ret
!= TRACE_TYPE_UNHANDLED
)
4054 if (iter
->ent
->type
== TRACE_BPUTS
&&
4055 trace_flags
& TRACE_ITER_PRINTK
&&
4056 trace_flags
& TRACE_ITER_PRINTK_MSGONLY
)
4057 return trace_print_bputs_msg_only(iter
);
4059 if (iter
->ent
->type
== TRACE_BPRINT
&&
4060 trace_flags
& TRACE_ITER_PRINTK
&&
4061 trace_flags
& TRACE_ITER_PRINTK_MSGONLY
)
4062 return trace_print_bprintk_msg_only(iter
);
4064 if (iter
->ent
->type
== TRACE_PRINT
&&
4065 trace_flags
& TRACE_ITER_PRINTK
&&
4066 trace_flags
& TRACE_ITER_PRINTK_MSGONLY
)
4067 return trace_print_printk_msg_only(iter
);
4069 if (trace_flags
& TRACE_ITER_BIN
)
4070 return print_bin_fmt(iter
);
4072 if (trace_flags
& TRACE_ITER_HEX
)
4073 return print_hex_fmt(iter
);
4075 if (trace_flags
& TRACE_ITER_RAW
)
4076 return print_raw_fmt(iter
);
4078 return print_trace_fmt(iter
);
4081 void trace_latency_header(struct seq_file
*m
)
4083 struct trace_iterator
*iter
= m
->private;
4084 struct trace_array
*tr
= iter
->tr
;
4086 /* print nothing if the buffers are empty */
4087 if (trace_empty(iter
))
4090 if (iter
->iter_flags
& TRACE_FILE_LAT_FMT
)
4091 print_trace_header(m
, iter
);
4093 if (!(tr
->trace_flags
& TRACE_ITER_VERBOSE
))
4094 print_lat_help_header(m
);
4097 void trace_default_header(struct seq_file
*m
)
4099 struct trace_iterator
*iter
= m
->private;
4100 struct trace_array
*tr
= iter
->tr
;
4101 unsigned long trace_flags
= tr
->trace_flags
;
4103 if (!(trace_flags
& TRACE_ITER_CONTEXT_INFO
))
4106 if (iter
->iter_flags
& TRACE_FILE_LAT_FMT
) {
4107 /* print nothing if the buffers are empty */
4108 if (trace_empty(iter
))
4110 print_trace_header(m
, iter
);
4111 if (!(trace_flags
& TRACE_ITER_VERBOSE
))
4112 print_lat_help_header(m
);
4114 if (!(trace_flags
& TRACE_ITER_VERBOSE
)) {
4115 if (trace_flags
& TRACE_ITER_IRQ_INFO
)
4116 print_func_help_header_irq(iter
->array_buffer
,
4119 print_func_help_header(iter
->array_buffer
, m
,
4125 static void test_ftrace_alive(struct seq_file
*m
)
4127 if (!ftrace_is_dead())
4129 seq_puts(m
, "# WARNING: FUNCTION TRACING IS CORRUPTED\n"
4130 "# MAY BE MISSING FUNCTION EVENTS\n");
4133 #ifdef CONFIG_TRACER_MAX_TRACE
4134 static void show_snapshot_main_help(struct seq_file
*m
)
4136 seq_puts(m
, "# echo 0 > snapshot : Clears and frees snapshot buffer\n"
4137 "# echo 1 > snapshot : Allocates snapshot buffer, if not already allocated.\n"
4138 "# Takes a snapshot of the main buffer.\n"
4139 "# echo 2 > snapshot : Clears snapshot buffer (but does not allocate or free)\n"
4140 "# (Doesn't have to be '2' works with any number that\n"
4141 "# is not a '0' or '1')\n");
4144 static void show_snapshot_percpu_help(struct seq_file
*m
)
4146 seq_puts(m
, "# echo 0 > snapshot : Invalid for per_cpu snapshot file.\n");
4147 #ifdef CONFIG_RING_BUFFER_ALLOW_SWAP
4148 seq_puts(m
, "# echo 1 > snapshot : Allocates snapshot buffer, if not already allocated.\n"
4149 "# Takes a snapshot of the main buffer for this cpu.\n");
4151 seq_puts(m
, "# echo 1 > snapshot : Not supported with this kernel.\n"
4152 "# Must use main snapshot file to allocate.\n");
4154 seq_puts(m
, "# echo 2 > snapshot : Clears this cpu's snapshot buffer (but does not allocate)\n"
4155 "# (Doesn't have to be '2' works with any number that\n"
4156 "# is not a '0' or '1')\n");
4159 static void print_snapshot_help(struct seq_file
*m
, struct trace_iterator
*iter
)
4161 if (iter
->tr
->allocated_snapshot
)
4162 seq_puts(m
, "#\n# * Snapshot is allocated *\n#\n");
4164 seq_puts(m
, "#\n# * Snapshot is freed *\n#\n");
4166 seq_puts(m
, "# Snapshot commands:\n");
4167 if (iter
->cpu_file
== RING_BUFFER_ALL_CPUS
)
4168 show_snapshot_main_help(m
);
4170 show_snapshot_percpu_help(m
);
4173 /* Should never be called */
4174 static inline void print_snapshot_help(struct seq_file
*m
, struct trace_iterator
*iter
) { }
4177 static int s_show(struct seq_file
*m
, void *v
)
4179 struct trace_iterator
*iter
= v
;
4182 if (iter
->ent
== NULL
) {
4184 seq_printf(m
, "# tracer: %s\n", iter
->trace
->name
);
4186 test_ftrace_alive(m
);
4188 if (iter
->snapshot
&& trace_empty(iter
))
4189 print_snapshot_help(m
, iter
);
4190 else if (iter
->trace
&& iter
->trace
->print_header
)
4191 iter
->trace
->print_header(m
);
4193 trace_default_header(m
);
4195 } else if (iter
->leftover
) {
4197 * If we filled the seq_file buffer earlier, we
4198 * want to just show it now.
4200 ret
= trace_print_seq(m
, &iter
->seq
);
4202 /* ret should this time be zero, but you never know */
4203 iter
->leftover
= ret
;
4206 print_trace_line(iter
);
4207 ret
= trace_print_seq(m
, &iter
->seq
);
4209 * If we overflow the seq_file buffer, then it will
4210 * ask us for this data again at start up.
4212 * ret is 0 if seq_file write succeeded.
4215 iter
->leftover
= ret
;
4222 * Should be used after trace_array_get(), trace_types_lock
4223 * ensures that i_cdev was already initialized.
4225 static inline int tracing_get_cpu(struct inode
*inode
)
4227 if (inode
->i_cdev
) /* See trace_create_cpu_file() */
4228 return (long)inode
->i_cdev
- 1;
4229 return RING_BUFFER_ALL_CPUS
;
4232 static const struct seq_operations tracer_seq_ops
= {
4239 static struct trace_iterator
*
4240 __tracing_open(struct inode
*inode
, struct file
*file
, bool snapshot
)
4242 struct trace_array
*tr
= inode
->i_private
;
4243 struct trace_iterator
*iter
;
4246 if (tracing_disabled
)
4247 return ERR_PTR(-ENODEV
);
4249 iter
= __seq_open_private(file
, &tracer_seq_ops
, sizeof(*iter
));
4251 return ERR_PTR(-ENOMEM
);
4253 iter
->buffer_iter
= kcalloc(nr_cpu_ids
, sizeof(*iter
->buffer_iter
),
4255 if (!iter
->buffer_iter
)
4259 * trace_find_next_entry() may need to save off iter->ent.
4260 * It will place it into the iter->temp buffer. As most
4261 * events are less than 128, allocate a buffer of that size.
4262 * If one is greater, then trace_find_next_entry() will
4263 * allocate a new buffer to adjust for the bigger iter->ent.
4264 * It's not critical if it fails to get allocated here.
4266 iter
->temp
= kmalloc(128, GFP_KERNEL
);
4268 iter
->temp_size
= 128;
4271 * We make a copy of the current tracer to avoid concurrent
4272 * changes on it while we are reading.
4274 mutex_lock(&trace_types_lock
);
4275 iter
->trace
= kzalloc(sizeof(*iter
->trace
), GFP_KERNEL
);
4279 *iter
->trace
= *tr
->current_trace
;
4281 if (!zalloc_cpumask_var(&iter
->started
, GFP_KERNEL
))
4286 #ifdef CONFIG_TRACER_MAX_TRACE
4287 /* Currently only the top directory has a snapshot */
4288 if (tr
->current_trace
->print_max
|| snapshot
)
4289 iter
->array_buffer
= &tr
->max_buffer
;
4292 iter
->array_buffer
= &tr
->array_buffer
;
4293 iter
->snapshot
= snapshot
;
4295 iter
->cpu_file
= tracing_get_cpu(inode
);
4296 mutex_init(&iter
->mutex
);
4298 /* Notify the tracer early; before we stop tracing. */
4299 if (iter
->trace
->open
)
4300 iter
->trace
->open(iter
);
4302 /* Annotate start of buffers if we had overruns */
4303 if (ring_buffer_overruns(iter
->array_buffer
->buffer
))
4304 iter
->iter_flags
|= TRACE_FILE_ANNOTATE
;
4306 /* Output in nanoseconds only if we are using a clock in nanoseconds. */
4307 if (trace_clocks
[tr
->clock_id
].in_ns
)
4308 iter
->iter_flags
|= TRACE_FILE_TIME_IN_NS
;
4311 * If pause-on-trace is enabled, then stop the trace while
4312 * dumping, unless this is the "snapshot" file
4314 if (!iter
->snapshot
&& (tr
->trace_flags
& TRACE_ITER_PAUSE_ON_TRACE
))
4315 tracing_stop_tr(tr
);
4317 if (iter
->cpu_file
== RING_BUFFER_ALL_CPUS
) {
4318 for_each_tracing_cpu(cpu
) {
4319 iter
->buffer_iter
[cpu
] =
4320 ring_buffer_read_prepare(iter
->array_buffer
->buffer
,
4323 ring_buffer_read_prepare_sync();
4324 for_each_tracing_cpu(cpu
) {
4325 ring_buffer_read_start(iter
->buffer_iter
[cpu
]);
4326 tracing_iter_reset(iter
, cpu
);
4329 cpu
= iter
->cpu_file
;
4330 iter
->buffer_iter
[cpu
] =
4331 ring_buffer_read_prepare(iter
->array_buffer
->buffer
,
4333 ring_buffer_read_prepare_sync();
4334 ring_buffer_read_start(iter
->buffer_iter
[cpu
]);
4335 tracing_iter_reset(iter
, cpu
);
4338 mutex_unlock(&trace_types_lock
);
4343 mutex_unlock(&trace_types_lock
);
4346 kfree(iter
->buffer_iter
);
4348 seq_release_private(inode
, file
);
4349 return ERR_PTR(-ENOMEM
);
4352 int tracing_open_generic(struct inode
*inode
, struct file
*filp
)
4356 ret
= tracing_check_open_get_tr(NULL
);
4360 filp
->private_data
= inode
->i_private
;
4364 bool tracing_is_disabled(void)
4366 return (tracing_disabled
) ? true: false;
4370 * Open and update trace_array ref count.
4371 * Must have the current trace_array passed to it.
4373 int tracing_open_generic_tr(struct inode
*inode
, struct file
*filp
)
4375 struct trace_array
*tr
= inode
->i_private
;
4378 ret
= tracing_check_open_get_tr(tr
);
4382 filp
->private_data
= inode
->i_private
;
4387 static int tracing_release(struct inode
*inode
, struct file
*file
)
4389 struct trace_array
*tr
= inode
->i_private
;
4390 struct seq_file
*m
= file
->private_data
;
4391 struct trace_iterator
*iter
;
4394 if (!(file
->f_mode
& FMODE_READ
)) {
4395 trace_array_put(tr
);
4399 /* Writes do not use seq_file */
4401 mutex_lock(&trace_types_lock
);
4403 for_each_tracing_cpu(cpu
) {
4404 if (iter
->buffer_iter
[cpu
])
4405 ring_buffer_read_finish(iter
->buffer_iter
[cpu
]);
4408 if (iter
->trace
&& iter
->trace
->close
)
4409 iter
->trace
->close(iter
);
4411 if (!iter
->snapshot
&& tr
->stop_count
)
4412 /* reenable tracing if it was previously enabled */
4413 tracing_start_tr(tr
);
4415 __trace_array_put(tr
);
4417 mutex_unlock(&trace_types_lock
);
4419 mutex_destroy(&iter
->mutex
);
4420 free_cpumask_var(iter
->started
);
4423 kfree(iter
->buffer_iter
);
4424 seq_release_private(inode
, file
);
4429 static int tracing_release_generic_tr(struct inode
*inode
, struct file
*file
)
4431 struct trace_array
*tr
= inode
->i_private
;
4433 trace_array_put(tr
);
4437 static int tracing_single_release_tr(struct inode
*inode
, struct file
*file
)
4439 struct trace_array
*tr
= inode
->i_private
;
4441 trace_array_put(tr
);
4443 return single_release(inode
, file
);
4446 static int tracing_open(struct inode
*inode
, struct file
*file
)
4448 struct trace_array
*tr
= inode
->i_private
;
4449 struct trace_iterator
*iter
;
4452 ret
= tracing_check_open_get_tr(tr
);
4456 /* If this file was open for write, then erase contents */
4457 if ((file
->f_mode
& FMODE_WRITE
) && (file
->f_flags
& O_TRUNC
)) {
4458 int cpu
= tracing_get_cpu(inode
);
4459 struct array_buffer
*trace_buf
= &tr
->array_buffer
;
4461 #ifdef CONFIG_TRACER_MAX_TRACE
4462 if (tr
->current_trace
->print_max
)
4463 trace_buf
= &tr
->max_buffer
;
4466 if (cpu
== RING_BUFFER_ALL_CPUS
)
4467 tracing_reset_online_cpus(trace_buf
);
4469 tracing_reset_cpu(trace_buf
, cpu
);
4472 if (file
->f_mode
& FMODE_READ
) {
4473 iter
= __tracing_open(inode
, file
, false);
4475 ret
= PTR_ERR(iter
);
4476 else if (tr
->trace_flags
& TRACE_ITER_LATENCY_FMT
)
4477 iter
->iter_flags
|= TRACE_FILE_LAT_FMT
;
4481 trace_array_put(tr
);
4487 * Some tracers are not suitable for instance buffers.
4488 * A tracer is always available for the global array (toplevel)
4489 * or if it explicitly states that it is.
4492 trace_ok_for_array(struct tracer
*t
, struct trace_array
*tr
)
4494 return (tr
->flags
& TRACE_ARRAY_FL_GLOBAL
) || t
->allow_instances
;
4497 /* Find the next tracer that this trace array may use */
4498 static struct tracer
*
4499 get_tracer_for_array(struct trace_array
*tr
, struct tracer
*t
)
4501 while (t
&& !trace_ok_for_array(t
, tr
))
4508 t_next(struct seq_file
*m
, void *v
, loff_t
*pos
)
4510 struct trace_array
*tr
= m
->private;
4511 struct tracer
*t
= v
;
4516 t
= get_tracer_for_array(tr
, t
->next
);
4521 static void *t_start(struct seq_file
*m
, loff_t
*pos
)
4523 struct trace_array
*tr
= m
->private;
4527 mutex_lock(&trace_types_lock
);
4529 t
= get_tracer_for_array(tr
, trace_types
);
4530 for (; t
&& l
< *pos
; t
= t_next(m
, t
, &l
))
4536 static void t_stop(struct seq_file
*m
, void *p
)
4538 mutex_unlock(&trace_types_lock
);
4541 static int t_show(struct seq_file
*m
, void *v
)
4543 struct tracer
*t
= v
;
4548 seq_puts(m
, t
->name
);
4557 static const struct seq_operations show_traces_seq_ops
= {
4564 static int show_traces_open(struct inode
*inode
, struct file
*file
)
4566 struct trace_array
*tr
= inode
->i_private
;
4570 ret
= tracing_check_open_get_tr(tr
);
4574 ret
= seq_open(file
, &show_traces_seq_ops
);
4576 trace_array_put(tr
);
4580 m
= file
->private_data
;
4586 static int show_traces_release(struct inode
*inode
, struct file
*file
)
4588 struct trace_array
*tr
= inode
->i_private
;
4590 trace_array_put(tr
);
4591 return seq_release(inode
, file
);
4595 tracing_write_stub(struct file
*filp
, const char __user
*ubuf
,
4596 size_t count
, loff_t
*ppos
)
4601 loff_t
tracing_lseek(struct file
*file
, loff_t offset
, int whence
)
4605 if (file
->f_mode
& FMODE_READ
)
4606 ret
= seq_lseek(file
, offset
, whence
);
4608 file
->f_pos
= ret
= 0;
4613 static const struct file_operations tracing_fops
= {
4614 .open
= tracing_open
,
4616 .write
= tracing_write_stub
,
4617 .llseek
= tracing_lseek
,
4618 .release
= tracing_release
,
4621 static const struct file_operations show_traces_fops
= {
4622 .open
= show_traces_open
,
4624 .llseek
= seq_lseek
,
4625 .release
= show_traces_release
,
4629 tracing_cpumask_read(struct file
*filp
, char __user
*ubuf
,
4630 size_t count
, loff_t
*ppos
)
4632 struct trace_array
*tr
= file_inode(filp
)->i_private
;
4636 len
= snprintf(NULL
, 0, "%*pb\n",
4637 cpumask_pr_args(tr
->tracing_cpumask
)) + 1;
4638 mask_str
= kmalloc(len
, GFP_KERNEL
);
4642 len
= snprintf(mask_str
, len
, "%*pb\n",
4643 cpumask_pr_args(tr
->tracing_cpumask
));
4648 count
= simple_read_from_buffer(ubuf
, count
, ppos
, mask_str
, len
);
4656 int tracing_set_cpumask(struct trace_array
*tr
,
4657 cpumask_var_t tracing_cpumask_new
)
4664 local_irq_disable();
4665 arch_spin_lock(&tr
->max_lock
);
4666 for_each_tracing_cpu(cpu
) {
4668 * Increase/decrease the disabled counter if we are
4669 * about to flip a bit in the cpumask:
4671 if (cpumask_test_cpu(cpu
, tr
->tracing_cpumask
) &&
4672 !cpumask_test_cpu(cpu
, tracing_cpumask_new
)) {
4673 atomic_inc(&per_cpu_ptr(tr
->array_buffer
.data
, cpu
)->disabled
);
4674 ring_buffer_record_disable_cpu(tr
->array_buffer
.buffer
, cpu
);
4676 if (!cpumask_test_cpu(cpu
, tr
->tracing_cpumask
) &&
4677 cpumask_test_cpu(cpu
, tracing_cpumask_new
)) {
4678 atomic_dec(&per_cpu_ptr(tr
->array_buffer
.data
, cpu
)->disabled
);
4679 ring_buffer_record_enable_cpu(tr
->array_buffer
.buffer
, cpu
);
4682 arch_spin_unlock(&tr
->max_lock
);
4685 cpumask_copy(tr
->tracing_cpumask
, tracing_cpumask_new
);
4691 tracing_cpumask_write(struct file
*filp
, const char __user
*ubuf
,
4692 size_t count
, loff_t
*ppos
)
4694 struct trace_array
*tr
= file_inode(filp
)->i_private
;
4695 cpumask_var_t tracing_cpumask_new
;
4698 if (!alloc_cpumask_var(&tracing_cpumask_new
, GFP_KERNEL
))
4701 err
= cpumask_parse_user(ubuf
, count
, tracing_cpumask_new
);
4705 err
= tracing_set_cpumask(tr
, tracing_cpumask_new
);
4709 free_cpumask_var(tracing_cpumask_new
);
4714 free_cpumask_var(tracing_cpumask_new
);
4719 static const struct file_operations tracing_cpumask_fops
= {
4720 .open
= tracing_open_generic_tr
,
4721 .read
= tracing_cpumask_read
,
4722 .write
= tracing_cpumask_write
,
4723 .release
= tracing_release_generic_tr
,
4724 .llseek
= generic_file_llseek
,
4727 static int tracing_trace_options_show(struct seq_file
*m
, void *v
)
4729 struct tracer_opt
*trace_opts
;
4730 struct trace_array
*tr
= m
->private;
4734 mutex_lock(&trace_types_lock
);
4735 tracer_flags
= tr
->current_trace
->flags
->val
;
4736 trace_opts
= tr
->current_trace
->flags
->opts
;
4738 for (i
= 0; trace_options
[i
]; i
++) {
4739 if (tr
->trace_flags
& (1 << i
))
4740 seq_printf(m
, "%s\n", trace_options
[i
]);
4742 seq_printf(m
, "no%s\n", trace_options
[i
]);
4745 for (i
= 0; trace_opts
[i
].name
; i
++) {
4746 if (tracer_flags
& trace_opts
[i
].bit
)
4747 seq_printf(m
, "%s\n", trace_opts
[i
].name
);
4749 seq_printf(m
, "no%s\n", trace_opts
[i
].name
);
4751 mutex_unlock(&trace_types_lock
);
4756 static int __set_tracer_option(struct trace_array
*tr
,
4757 struct tracer_flags
*tracer_flags
,
4758 struct tracer_opt
*opts
, int neg
)
4760 struct tracer
*trace
= tracer_flags
->trace
;
4763 ret
= trace
->set_flag(tr
, tracer_flags
->val
, opts
->bit
, !neg
);
4768 tracer_flags
->val
&= ~opts
->bit
;
4770 tracer_flags
->val
|= opts
->bit
;
4774 /* Try to assign a tracer specific option */
4775 static int set_tracer_option(struct trace_array
*tr
, char *cmp
, int neg
)
4777 struct tracer
*trace
= tr
->current_trace
;
4778 struct tracer_flags
*tracer_flags
= trace
->flags
;
4779 struct tracer_opt
*opts
= NULL
;
4782 for (i
= 0; tracer_flags
->opts
[i
].name
; i
++) {
4783 opts
= &tracer_flags
->opts
[i
];
4785 if (strcmp(cmp
, opts
->name
) == 0)
4786 return __set_tracer_option(tr
, trace
->flags
, opts
, neg
);
4792 /* Some tracers require overwrite to stay enabled */
4793 int trace_keep_overwrite(struct tracer
*tracer
, u32 mask
, int set
)
4795 if (tracer
->enabled
&& (mask
& TRACE_ITER_OVERWRITE
) && !set
)
4801 int set_tracer_flag(struct trace_array
*tr
, unsigned int mask
, int enabled
)
4803 if ((mask
== TRACE_ITER_RECORD_TGID
) ||
4804 (mask
== TRACE_ITER_RECORD_CMD
))
4805 lockdep_assert_held(&event_mutex
);
4807 /* do nothing if flag is already set */
4808 if (!!(tr
->trace_flags
& mask
) == !!enabled
)
4811 /* Give the tracer a chance to approve the change */
4812 if (tr
->current_trace
->flag_changed
)
4813 if (tr
->current_trace
->flag_changed(tr
, mask
, !!enabled
))
4817 tr
->trace_flags
|= mask
;
4819 tr
->trace_flags
&= ~mask
;
4821 if (mask
== TRACE_ITER_RECORD_CMD
)
4822 trace_event_enable_cmd_record(enabled
);
4824 if (mask
== TRACE_ITER_RECORD_TGID
) {
4826 tgid_map
= kvcalloc(PID_MAX_DEFAULT
+ 1,
4830 tr
->trace_flags
&= ~TRACE_ITER_RECORD_TGID
;
4834 trace_event_enable_tgid_record(enabled
);
4837 if (mask
== TRACE_ITER_EVENT_FORK
)
4838 trace_event_follow_fork(tr
, enabled
);
4840 if (mask
== TRACE_ITER_FUNC_FORK
)
4841 ftrace_pid_follow_fork(tr
, enabled
);
4843 if (mask
== TRACE_ITER_OVERWRITE
) {
4844 ring_buffer_change_overwrite(tr
->array_buffer
.buffer
, enabled
);
4845 #ifdef CONFIG_TRACER_MAX_TRACE
4846 ring_buffer_change_overwrite(tr
->max_buffer
.buffer
, enabled
);
4850 if (mask
== TRACE_ITER_PRINTK
) {
4851 trace_printk_start_stop_comm(enabled
);
4852 trace_printk_control(enabled
);
4858 int trace_set_options(struct trace_array
*tr
, char *option
)
4863 size_t orig_len
= strlen(option
);
4866 cmp
= strstrip(option
);
4868 len
= str_has_prefix(cmp
, "no");
4874 mutex_lock(&event_mutex
);
4875 mutex_lock(&trace_types_lock
);
4877 ret
= match_string(trace_options
, -1, cmp
);
4878 /* If no option could be set, test the specific tracer options */
4880 ret
= set_tracer_option(tr
, cmp
, neg
);
4882 ret
= set_tracer_flag(tr
, 1 << ret
, !neg
);
4884 mutex_unlock(&trace_types_lock
);
4885 mutex_unlock(&event_mutex
);
4888 * If the first trailing whitespace is replaced with '\0' by strstrip,
4889 * turn it back into a space.
4891 if (orig_len
> strlen(option
))
4892 option
[strlen(option
)] = ' ';
4897 static void __init
apply_trace_boot_options(void)
4899 char *buf
= trace_boot_options_buf
;
4903 option
= strsep(&buf
, ",");
4909 trace_set_options(&global_trace
, option
);
4911 /* Put back the comma to allow this to be called again */
4918 tracing_trace_options_write(struct file
*filp
, const char __user
*ubuf
,
4919 size_t cnt
, loff_t
*ppos
)
4921 struct seq_file
*m
= filp
->private_data
;
4922 struct trace_array
*tr
= m
->private;
4926 if (cnt
>= sizeof(buf
))
4929 if (copy_from_user(buf
, ubuf
, cnt
))
4934 ret
= trace_set_options(tr
, buf
);
4943 static int tracing_trace_options_open(struct inode
*inode
, struct file
*file
)
4945 struct trace_array
*tr
= inode
->i_private
;
4948 ret
= tracing_check_open_get_tr(tr
);
4952 ret
= single_open(file
, tracing_trace_options_show
, inode
->i_private
);
4954 trace_array_put(tr
);
4959 static const struct file_operations tracing_iter_fops
= {
4960 .open
= tracing_trace_options_open
,
4962 .llseek
= seq_lseek
,
4963 .release
= tracing_single_release_tr
,
4964 .write
= tracing_trace_options_write
,
4967 static const char readme_msg
[] =
4968 "tracing mini-HOWTO:\n\n"
4969 "# echo 0 > tracing_on : quick way to disable tracing\n"
4970 "# echo 1 > tracing_on : quick way to re-enable tracing\n\n"
4971 " Important files:\n"
4972 " trace\t\t\t- The static contents of the buffer\n"
4973 "\t\t\t To clear the buffer write into this file: echo > trace\n"
4974 " trace_pipe\t\t- A consuming read to see the contents of the buffer\n"
4975 " current_tracer\t- function and latency tracers\n"
4976 " available_tracers\t- list of configured tracers for current_tracer\n"
4977 " error_log\t- error log for failed commands (that support it)\n"
4978 " buffer_size_kb\t- view and modify size of per cpu buffer\n"
4979 " buffer_total_size_kb - view total size of all cpu buffers\n\n"
4980 " trace_clock\t\t-change the clock used to order events\n"
4981 " local: Per cpu clock but may not be synced across CPUs\n"
4982 " global: Synced across CPUs but slows tracing down.\n"
4983 " counter: Not a clock, but just an increment\n"
4984 " uptime: Jiffy counter from time of boot\n"
4985 " perf: Same clock that perf events use\n"
4986 #ifdef CONFIG_X86_64
4987 " x86-tsc: TSC cycle counter\n"
4989 "\n timestamp_mode\t-view the mode used to timestamp events\n"
4990 " delta: Delta difference against a buffer-wide timestamp\n"
4991 " absolute: Absolute (standalone) timestamp\n"
4992 "\n trace_marker\t\t- Writes into this file writes into the kernel buffer\n"
4993 "\n trace_marker_raw\t\t- Writes into this file writes binary data into the kernel buffer\n"
4994 " tracing_cpumask\t- Limit which CPUs to trace\n"
4995 " instances\t\t- Make sub-buffers with: mkdir instances/foo\n"
4996 "\t\t\t Remove sub-buffer with rmdir\n"
4997 " trace_options\t\t- Set format or modify how tracing happens\n"
4998 "\t\t\t Disable an option by prefixing 'no' to the\n"
4999 "\t\t\t option name\n"
5000 " saved_cmdlines_size\t- echo command number in here to store comm-pid list\n"
5001 #ifdef CONFIG_DYNAMIC_FTRACE
5002 "\n available_filter_functions - list of functions that can be filtered on\n"
5003 " set_ftrace_filter\t- echo function name in here to only trace these\n"
5004 "\t\t\t functions\n"
5005 "\t accepts: func_full_name or glob-matching-pattern\n"
5006 "\t modules: Can select a group via module\n"
5007 "\t Format: :mod:<module-name>\n"
5008 "\t example: echo :mod:ext3 > set_ftrace_filter\n"
5009 "\t triggers: a command to perform when function is hit\n"
5010 "\t Format: <function>:<trigger>[:count]\n"
5011 "\t trigger: traceon, traceoff\n"
5012 "\t\t enable_event:<system>:<event>\n"
5013 "\t\t disable_event:<system>:<event>\n"
5014 #ifdef CONFIG_STACKTRACE
5017 #ifdef CONFIG_TRACER_SNAPSHOT
5022 "\t example: echo do_fault:traceoff > set_ftrace_filter\n"
5023 "\t echo do_trap:traceoff:3 > set_ftrace_filter\n"
5024 "\t The first one will disable tracing every time do_fault is hit\n"
5025 "\t The second will disable tracing at most 3 times when do_trap is hit\n"
5026 "\t The first time do trap is hit and it disables tracing, the\n"
5027 "\t counter will decrement to 2. If tracing is already disabled,\n"
5028 "\t the counter will not decrement. It only decrements when the\n"
5029 "\t trigger did work\n"
5030 "\t To remove trigger without count:\n"
5031 "\t echo '!<function>:<trigger> > set_ftrace_filter\n"
5032 "\t To remove trigger with a count:\n"
5033 "\t echo '!<function>:<trigger>:0 > set_ftrace_filter\n"
5034 " set_ftrace_notrace\t- echo function name in here to never trace.\n"
5035 "\t accepts: func_full_name, *func_end, func_begin*, *func_middle*\n"
5036 "\t modules: Can select a group via module command :mod:\n"
5037 "\t Does not accept triggers\n"
5038 #endif /* CONFIG_DYNAMIC_FTRACE */
5039 #ifdef CONFIG_FUNCTION_TRACER
5040 " set_ftrace_pid\t- Write pid(s) to only function trace those pids\n"
5042 " set_ftrace_notrace_pid\t- Write pid(s) to not function trace those pids\n"
5045 #ifdef CONFIG_FUNCTION_GRAPH_TRACER
5046 " set_graph_function\t- Trace the nested calls of a function (function_graph)\n"
5047 " set_graph_notrace\t- Do not trace the nested calls of a function (function_graph)\n"
5048 " max_graph_depth\t- Trace a limited depth of nested calls (0 is unlimited)\n"
5050 #ifdef CONFIG_TRACER_SNAPSHOT
5051 "\n snapshot\t\t- Like 'trace' but shows the content of the static\n"
5052 "\t\t\t snapshot buffer. Read the contents for more\n"
5053 "\t\t\t information\n"
5055 #ifdef CONFIG_STACK_TRACER
5056 " stack_trace\t\t- Shows the max stack trace when active\n"
5057 " stack_max_size\t- Shows current max stack size that was traced\n"
5058 "\t\t\t Write into this file to reset the max size (trigger a\n"
5059 "\t\t\t new trace)\n"
5060 #ifdef CONFIG_DYNAMIC_FTRACE
5061 " stack_trace_filter\t- Like set_ftrace_filter but limits what stack_trace\n"
5064 #endif /* CONFIG_STACK_TRACER */
5065 #ifdef CONFIG_DYNAMIC_EVENTS
5066 " dynamic_events\t\t- Create/append/remove/show the generic dynamic events\n"
5067 "\t\t\t Write into this file to define/undefine new trace events.\n"
5069 #ifdef CONFIG_KPROBE_EVENTS
5070 " kprobe_events\t\t- Create/append/remove/show the kernel dynamic events\n"
5071 "\t\t\t Write into this file to define/undefine new trace events.\n"
5073 #ifdef CONFIG_UPROBE_EVENTS
5074 " uprobe_events\t\t- Create/append/remove/show the userspace dynamic events\n"
5075 "\t\t\t Write into this file to define/undefine new trace events.\n"
5077 #if defined(CONFIG_KPROBE_EVENTS) || defined(CONFIG_UPROBE_EVENTS)
5078 "\t accepts: event-definitions (one definition per line)\n"
5079 "\t Format: p[:[<group>/]<event>] <place> [<args>]\n"
5080 "\t r[maxactive][:[<group>/]<event>] <place> [<args>]\n"
5081 #ifdef CONFIG_HIST_TRIGGERS
5082 "\t s:[synthetic/]<event> <field> [<field>]\n"
5084 "\t -:[<group>/]<event>\n"
5085 #ifdef CONFIG_KPROBE_EVENTS
5086 "\t place: [<module>:]<symbol>[+<offset>]|<memaddr>\n"
5087 "place (kretprobe): [<module>:]<symbol>[+<offset>]|<memaddr>\n"
5089 #ifdef CONFIG_UPROBE_EVENTS
5090 " place (uprobe): <path>:<offset>[(ref_ctr_offset)]\n"
5092 "\t args: <name>=fetcharg[:type]\n"
5093 "\t fetcharg: %<register>, @<address>, @<symbol>[+|-<offset>],\n"
5094 #ifdef CONFIG_HAVE_FUNCTION_ARG_ACCESS_API
5095 "\t $stack<index>, $stack, $retval, $comm, $arg<N>,\n"
5097 "\t $stack<index>, $stack, $retval, $comm,\n"
5099 "\t +|-[u]<offset>(<fetcharg>), \\imm-value, \\\"imm-string\"\n"
5100 "\t type: s8/16/32/64, u8/16/32/64, x8/16/32/64, string, symbol,\n"
5101 "\t b<bit-width>@<bit-offset>/<container-size>, ustring,\n"
5102 "\t <type>\\[<array-size>\\]\n"
5103 #ifdef CONFIG_HIST_TRIGGERS
5104 "\t field: <stype> <name>;\n"
5105 "\t stype: u8/u16/u32/u64, s8/s16/s32/s64, pid_t,\n"
5106 "\t [unsigned] char/int/long\n"
5109 " events/\t\t- Directory containing all trace event subsystems:\n"
5110 " enable\t\t- Write 0/1 to enable/disable tracing of all events\n"
5111 " events/<system>/\t- Directory containing all trace events for <system>:\n"
5112 " enable\t\t- Write 0/1 to enable/disable tracing of all <system>\n"
5114 " filter\t\t- If set, only events passing filter are traced\n"
5115 " events/<system>/<event>/\t- Directory containing control files for\n"
5117 " enable\t\t- Write 0/1 to enable/disable tracing of <event>\n"
5118 " filter\t\t- If set, only events passing filter are traced\n"
5119 " trigger\t\t- If set, a command to perform when event is hit\n"
5120 "\t Format: <trigger>[:count][if <filter>]\n"
5121 "\t trigger: traceon, traceoff\n"
5122 "\t enable_event:<system>:<event>\n"
5123 "\t disable_event:<system>:<event>\n"
5124 #ifdef CONFIG_HIST_TRIGGERS
5125 "\t enable_hist:<system>:<event>\n"
5126 "\t disable_hist:<system>:<event>\n"
5128 #ifdef CONFIG_STACKTRACE
5131 #ifdef CONFIG_TRACER_SNAPSHOT
5134 #ifdef CONFIG_HIST_TRIGGERS
5135 "\t\t hist (see below)\n"
5137 "\t example: echo traceoff > events/block/block_unplug/trigger\n"
5138 "\t echo traceoff:3 > events/block/block_unplug/trigger\n"
5139 "\t echo 'enable_event:kmem:kmalloc:3 if nr_rq > 1' > \\\n"
5140 "\t events/block/block_unplug/trigger\n"
5141 "\t The first disables tracing every time block_unplug is hit.\n"
5142 "\t The second disables tracing the first 3 times block_unplug is hit.\n"
5143 "\t The third enables the kmalloc event the first 3 times block_unplug\n"
5144 "\t is hit and has value of greater than 1 for the 'nr_rq' event field.\n"
5145 "\t Like function triggers, the counter is only decremented if it\n"
5146 "\t enabled or disabled tracing.\n"
5147 "\t To remove a trigger without a count:\n"
5148 "\t echo '!<trigger> > <system>/<event>/trigger\n"
5149 "\t To remove a trigger with a count:\n"
5150 "\t echo '!<trigger>:0 > <system>/<event>/trigger\n"
5151 "\t Filters can be ignored when removing a trigger.\n"
5152 #ifdef CONFIG_HIST_TRIGGERS
5153 " hist trigger\t- If set, event hits are aggregated into a hash table\n"
5154 "\t Format: hist:keys=<field1[,field2,...]>\n"
5155 "\t [:values=<field1[,field2,...]>]\n"
5156 "\t [:sort=<field1[,field2,...]>]\n"
5157 "\t [:size=#entries]\n"
5158 "\t [:pause][:continue][:clear]\n"
5159 "\t [:name=histname1]\n"
5160 "\t [:<handler>.<action>]\n"
5161 "\t [if <filter>]\n\n"
5162 "\t When a matching event is hit, an entry is added to a hash\n"
5163 "\t table using the key(s) and value(s) named, and the value of a\n"
5164 "\t sum called 'hitcount' is incremented. Keys and values\n"
5165 "\t correspond to fields in the event's format description. Keys\n"
5166 "\t can be any field, or the special string 'stacktrace'.\n"
5167 "\t Compound keys consisting of up to two fields can be specified\n"
5168 "\t by the 'keys' keyword. Values must correspond to numeric\n"
5169 "\t fields. Sort keys consisting of up to two fields can be\n"
5170 "\t specified using the 'sort' keyword. The sort direction can\n"
5171 "\t be modified by appending '.descending' or '.ascending' to a\n"
5172 "\t sort field. The 'size' parameter can be used to specify more\n"
5173 "\t or fewer than the default 2048 entries for the hashtable size.\n"
5174 "\t If a hist trigger is given a name using the 'name' parameter,\n"
5175 "\t its histogram data will be shared with other triggers of the\n"
5176 "\t same name, and trigger hits will update this common data.\n\n"
5177 "\t Reading the 'hist' file for the event will dump the hash\n"
5178 "\t table in its entirety to stdout. If there are multiple hist\n"
5179 "\t triggers attached to an event, there will be a table for each\n"
5180 "\t trigger in the output. The table displayed for a named\n"
5181 "\t trigger will be the same as any other instance having the\n"
5182 "\t same name. The default format used to display a given field\n"
5183 "\t can be modified by appending any of the following modifiers\n"
5184 "\t to the field name, as applicable:\n\n"
5185 "\t .hex display a number as a hex value\n"
5186 "\t .sym display an address as a symbol\n"
5187 "\t .sym-offset display an address as a symbol and offset\n"
5188 "\t .execname display a common_pid as a program name\n"
5189 "\t .syscall display a syscall id as a syscall name\n"
5190 "\t .log2 display log2 value rather than raw number\n"
5191 "\t .usecs display a common_timestamp in microseconds\n\n"
5192 "\t The 'pause' parameter can be used to pause an existing hist\n"
5193 "\t trigger or to start a hist trigger but not log any events\n"
5194 "\t until told to do so. 'continue' can be used to start or\n"
5195 "\t restart a paused hist trigger.\n\n"
5196 "\t The 'clear' parameter will clear the contents of a running\n"
5197 "\t hist trigger and leave its current paused/active state\n"
5199 "\t The enable_hist and disable_hist triggers can be used to\n"
5200 "\t have one event conditionally start and stop another event's\n"
5201 "\t already-attached hist trigger. The syntax is analogous to\n"
5202 "\t the enable_event and disable_event triggers.\n\n"
5203 "\t Hist trigger handlers and actions are executed whenever a\n"
5204 "\t a histogram entry is added or updated. They take the form:\n\n"
5205 "\t <handler>.<action>\n\n"
5206 "\t The available handlers are:\n\n"
5207 "\t onmatch(matching.event) - invoke on addition or update\n"
5208 "\t onmax(var) - invoke if var exceeds current max\n"
5209 "\t onchange(var) - invoke action if var changes\n\n"
5210 "\t The available actions are:\n\n"
5211 "\t trace(<synthetic_event>,param list) - generate synthetic event\n"
5212 "\t save(field,...) - save current event fields\n"
5213 #ifdef CONFIG_TRACER_SNAPSHOT
5214 "\t snapshot() - snapshot the trace buffer\n"
5220 tracing_readme_read(struct file
*filp
, char __user
*ubuf
,
5221 size_t cnt
, loff_t
*ppos
)
5223 return simple_read_from_buffer(ubuf
, cnt
, ppos
,
5224 readme_msg
, strlen(readme_msg
));
5227 static const struct file_operations tracing_readme_fops
= {
5228 .open
= tracing_open_generic
,
5229 .read
= tracing_readme_read
,
5230 .llseek
= generic_file_llseek
,
5233 static void *saved_tgids_next(struct seq_file
*m
, void *v
, loff_t
*pos
)
5237 if (*pos
|| m
->count
)
5242 for (; ptr
<= &tgid_map
[PID_MAX_DEFAULT
]; ptr
++) {
5243 if (trace_find_tgid(*ptr
))
5250 static void *saved_tgids_start(struct seq_file
*m
, loff_t
*pos
)
5260 v
= saved_tgids_next(m
, v
, &l
);
5268 static void saved_tgids_stop(struct seq_file
*m
, void *v
)
5272 static int saved_tgids_show(struct seq_file
*m
, void *v
)
5274 int pid
= (int *)v
- tgid_map
;
5276 seq_printf(m
, "%d %d\n", pid
, trace_find_tgid(pid
));
5280 static const struct seq_operations tracing_saved_tgids_seq_ops
= {
5281 .start
= saved_tgids_start
,
5282 .stop
= saved_tgids_stop
,
5283 .next
= saved_tgids_next
,
5284 .show
= saved_tgids_show
,
5287 static int tracing_saved_tgids_open(struct inode
*inode
, struct file
*filp
)
5291 ret
= tracing_check_open_get_tr(NULL
);
5295 return seq_open(filp
, &tracing_saved_tgids_seq_ops
);
5299 static const struct file_operations tracing_saved_tgids_fops
= {
5300 .open
= tracing_saved_tgids_open
,
5302 .llseek
= seq_lseek
,
5303 .release
= seq_release
,
5306 static void *saved_cmdlines_next(struct seq_file
*m
, void *v
, loff_t
*pos
)
5308 unsigned int *ptr
= v
;
5310 if (*pos
|| m
->count
)
5315 for (; ptr
< &savedcmd
->map_cmdline_to_pid
[savedcmd
->cmdline_num
];
5317 if (*ptr
== -1 || *ptr
== NO_CMDLINE_MAP
)
5326 static void *saved_cmdlines_start(struct seq_file
*m
, loff_t
*pos
)
5332 arch_spin_lock(&trace_cmdline_lock
);
5334 v
= &savedcmd
->map_cmdline_to_pid
[0];
5336 v
= saved_cmdlines_next(m
, v
, &l
);
5344 static void saved_cmdlines_stop(struct seq_file
*m
, void *v
)
5346 arch_spin_unlock(&trace_cmdline_lock
);
5350 static int saved_cmdlines_show(struct seq_file
*m
, void *v
)
5352 char buf
[TASK_COMM_LEN
];
5353 unsigned int *pid
= v
;
5355 __trace_find_cmdline(*pid
, buf
);
5356 seq_printf(m
, "%d %s\n", *pid
, buf
);
5360 static const struct seq_operations tracing_saved_cmdlines_seq_ops
= {
5361 .start
= saved_cmdlines_start
,
5362 .next
= saved_cmdlines_next
,
5363 .stop
= saved_cmdlines_stop
,
5364 .show
= saved_cmdlines_show
,
5367 static int tracing_saved_cmdlines_open(struct inode
*inode
, struct file
*filp
)
5371 ret
= tracing_check_open_get_tr(NULL
);
5375 return seq_open(filp
, &tracing_saved_cmdlines_seq_ops
);
5378 static const struct file_operations tracing_saved_cmdlines_fops
= {
5379 .open
= tracing_saved_cmdlines_open
,
5381 .llseek
= seq_lseek
,
5382 .release
= seq_release
,
5386 tracing_saved_cmdlines_size_read(struct file
*filp
, char __user
*ubuf
,
5387 size_t cnt
, loff_t
*ppos
)
5392 arch_spin_lock(&trace_cmdline_lock
);
5393 r
= scnprintf(buf
, sizeof(buf
), "%u\n", savedcmd
->cmdline_num
);
5394 arch_spin_unlock(&trace_cmdline_lock
);
5396 return simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, r
);
5399 static void free_saved_cmdlines_buffer(struct saved_cmdlines_buffer
*s
)
5401 kfree(s
->saved_cmdlines
);
5402 kfree(s
->map_cmdline_to_pid
);
5406 static int tracing_resize_saved_cmdlines(unsigned int val
)
5408 struct saved_cmdlines_buffer
*s
, *savedcmd_temp
;
5410 s
= kmalloc(sizeof(*s
), GFP_KERNEL
);
5414 if (allocate_cmdlines_buffer(val
, s
) < 0) {
5419 arch_spin_lock(&trace_cmdline_lock
);
5420 savedcmd_temp
= savedcmd
;
5422 arch_spin_unlock(&trace_cmdline_lock
);
5423 free_saved_cmdlines_buffer(savedcmd_temp
);
5429 tracing_saved_cmdlines_size_write(struct file
*filp
, const char __user
*ubuf
,
5430 size_t cnt
, loff_t
*ppos
)
5435 ret
= kstrtoul_from_user(ubuf
, cnt
, 10, &val
);
5439 /* must have at least 1 entry or less than PID_MAX_DEFAULT */
5440 if (!val
|| val
> PID_MAX_DEFAULT
)
5443 ret
= tracing_resize_saved_cmdlines((unsigned int)val
);
5452 static const struct file_operations tracing_saved_cmdlines_size_fops
= {
5453 .open
= tracing_open_generic
,
5454 .read
= tracing_saved_cmdlines_size_read
,
5455 .write
= tracing_saved_cmdlines_size_write
,
5458 #ifdef CONFIG_TRACE_EVAL_MAP_FILE
5459 static union trace_eval_map_item
*
5460 update_eval_map(union trace_eval_map_item
*ptr
)
5462 if (!ptr
->map
.eval_string
) {
5463 if (ptr
->tail
.next
) {
5464 ptr
= ptr
->tail
.next
;
5465 /* Set ptr to the next real item (skip head) */
5473 static void *eval_map_next(struct seq_file
*m
, void *v
, loff_t
*pos
)
5475 union trace_eval_map_item
*ptr
= v
;
5478 * Paranoid! If ptr points to end, we don't want to increment past it.
5479 * This really should never happen.
5482 ptr
= update_eval_map(ptr
);
5483 if (WARN_ON_ONCE(!ptr
))
5487 ptr
= update_eval_map(ptr
);
5492 static void *eval_map_start(struct seq_file
*m
, loff_t
*pos
)
5494 union trace_eval_map_item
*v
;
5497 mutex_lock(&trace_eval_mutex
);
5499 v
= trace_eval_maps
;
5503 while (v
&& l
< *pos
) {
5504 v
= eval_map_next(m
, v
, &l
);
5510 static void eval_map_stop(struct seq_file
*m
, void *v
)
5512 mutex_unlock(&trace_eval_mutex
);
5515 static int eval_map_show(struct seq_file
*m
, void *v
)
5517 union trace_eval_map_item
*ptr
= v
;
5519 seq_printf(m
, "%s %ld (%s)\n",
5520 ptr
->map
.eval_string
, ptr
->map
.eval_value
,
5526 static const struct seq_operations tracing_eval_map_seq_ops
= {
5527 .start
= eval_map_start
,
5528 .next
= eval_map_next
,
5529 .stop
= eval_map_stop
,
5530 .show
= eval_map_show
,
5533 static int tracing_eval_map_open(struct inode
*inode
, struct file
*filp
)
5537 ret
= tracing_check_open_get_tr(NULL
);
5541 return seq_open(filp
, &tracing_eval_map_seq_ops
);
5544 static const struct file_operations tracing_eval_map_fops
= {
5545 .open
= tracing_eval_map_open
,
5547 .llseek
= seq_lseek
,
5548 .release
= seq_release
,
5551 static inline union trace_eval_map_item
*
5552 trace_eval_jmp_to_tail(union trace_eval_map_item
*ptr
)
5554 /* Return tail of array given the head */
5555 return ptr
+ ptr
->head
.length
+ 1;
5559 trace_insert_eval_map_file(struct module
*mod
, struct trace_eval_map
**start
,
5562 struct trace_eval_map
**stop
;
5563 struct trace_eval_map
**map
;
5564 union trace_eval_map_item
*map_array
;
5565 union trace_eval_map_item
*ptr
;
5570 * The trace_eval_maps contains the map plus a head and tail item,
5571 * where the head holds the module and length of array, and the
5572 * tail holds a pointer to the next list.
5574 map_array
= kmalloc_array(len
+ 2, sizeof(*map_array
), GFP_KERNEL
);
5576 pr_warn("Unable to allocate trace eval mapping\n");
5580 mutex_lock(&trace_eval_mutex
);
5582 if (!trace_eval_maps
)
5583 trace_eval_maps
= map_array
;
5585 ptr
= trace_eval_maps
;
5587 ptr
= trace_eval_jmp_to_tail(ptr
);
5588 if (!ptr
->tail
.next
)
5590 ptr
= ptr
->tail
.next
;
5593 ptr
->tail
.next
= map_array
;
5595 map_array
->head
.mod
= mod
;
5596 map_array
->head
.length
= len
;
5599 for (map
= start
; (unsigned long)map
< (unsigned long)stop
; map
++) {
5600 map_array
->map
= **map
;
5603 memset(map_array
, 0, sizeof(*map_array
));
5605 mutex_unlock(&trace_eval_mutex
);
5608 static void trace_create_eval_file(struct dentry
*d_tracer
)
5610 trace_create_file("eval_map", 0444, d_tracer
,
5611 NULL
, &tracing_eval_map_fops
);
5614 #else /* CONFIG_TRACE_EVAL_MAP_FILE */
5615 static inline void trace_create_eval_file(struct dentry
*d_tracer
) { }
5616 static inline void trace_insert_eval_map_file(struct module
*mod
,
5617 struct trace_eval_map
**start
, int len
) { }
5618 #endif /* !CONFIG_TRACE_EVAL_MAP_FILE */
5620 static void trace_insert_eval_map(struct module
*mod
,
5621 struct trace_eval_map
**start
, int len
)
5623 struct trace_eval_map
**map
;
5630 trace_event_eval_update(map
, len
);
5632 trace_insert_eval_map_file(mod
, start
, len
);
5636 tracing_set_trace_read(struct file
*filp
, char __user
*ubuf
,
5637 size_t cnt
, loff_t
*ppos
)
5639 struct trace_array
*tr
= filp
->private_data
;
5640 char buf
[MAX_TRACER_SIZE
+2];
5643 mutex_lock(&trace_types_lock
);
5644 r
= sprintf(buf
, "%s\n", tr
->current_trace
->name
);
5645 mutex_unlock(&trace_types_lock
);
5647 return simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, r
);
5650 int tracer_init(struct tracer
*t
, struct trace_array
*tr
)
5652 tracing_reset_online_cpus(&tr
->array_buffer
);
5656 static void set_buffer_entries(struct array_buffer
*buf
, unsigned long val
)
5660 for_each_tracing_cpu(cpu
)
5661 per_cpu_ptr(buf
->data
, cpu
)->entries
= val
;
5664 #ifdef CONFIG_TRACER_MAX_TRACE
5665 /* resize @tr's buffer to the size of @size_tr's entries */
5666 static int resize_buffer_duplicate_size(struct array_buffer
*trace_buf
,
5667 struct array_buffer
*size_buf
, int cpu_id
)
5671 if (cpu_id
== RING_BUFFER_ALL_CPUS
) {
5672 for_each_tracing_cpu(cpu
) {
5673 ret
= ring_buffer_resize(trace_buf
->buffer
,
5674 per_cpu_ptr(size_buf
->data
, cpu
)->entries
, cpu
);
5677 per_cpu_ptr(trace_buf
->data
, cpu
)->entries
=
5678 per_cpu_ptr(size_buf
->data
, cpu
)->entries
;
5681 ret
= ring_buffer_resize(trace_buf
->buffer
,
5682 per_cpu_ptr(size_buf
->data
, cpu_id
)->entries
, cpu_id
);
5684 per_cpu_ptr(trace_buf
->data
, cpu_id
)->entries
=
5685 per_cpu_ptr(size_buf
->data
, cpu_id
)->entries
;
5690 #endif /* CONFIG_TRACER_MAX_TRACE */
5692 static int __tracing_resize_ring_buffer(struct trace_array
*tr
,
5693 unsigned long size
, int cpu
)
5698 * If kernel or user changes the size of the ring buffer
5699 * we use the size that was given, and we can forget about
5700 * expanding it later.
5702 ring_buffer_expanded
= true;
5704 /* May be called before buffers are initialized */
5705 if (!tr
->array_buffer
.buffer
)
5708 ret
= ring_buffer_resize(tr
->array_buffer
.buffer
, size
, cpu
);
5712 #ifdef CONFIG_TRACER_MAX_TRACE
5713 if (!(tr
->flags
& TRACE_ARRAY_FL_GLOBAL
) ||
5714 !tr
->current_trace
->use_max_tr
)
5717 ret
= ring_buffer_resize(tr
->max_buffer
.buffer
, size
, cpu
);
5719 int r
= resize_buffer_duplicate_size(&tr
->array_buffer
,
5720 &tr
->array_buffer
, cpu
);
5723 * AARGH! We are left with different
5724 * size max buffer!!!!
5725 * The max buffer is our "snapshot" buffer.
5726 * When a tracer needs a snapshot (one of the
5727 * latency tracers), it swaps the max buffer
5728 * with the saved snap shot. We succeeded to
5729 * update the size of the main buffer, but failed to
5730 * update the size of the max buffer. But when we tried
5731 * to reset the main buffer to the original size, we
5732 * failed there too. This is very unlikely to
5733 * happen, but if it does, warn and kill all
5737 tracing_disabled
= 1;
5742 if (cpu
== RING_BUFFER_ALL_CPUS
)
5743 set_buffer_entries(&tr
->max_buffer
, size
);
5745 per_cpu_ptr(tr
->max_buffer
.data
, cpu
)->entries
= size
;
5748 #endif /* CONFIG_TRACER_MAX_TRACE */
5750 if (cpu
== RING_BUFFER_ALL_CPUS
)
5751 set_buffer_entries(&tr
->array_buffer
, size
);
5753 per_cpu_ptr(tr
->array_buffer
.data
, cpu
)->entries
= size
;
5758 ssize_t
tracing_resize_ring_buffer(struct trace_array
*tr
,
5759 unsigned long size
, int cpu_id
)
5763 mutex_lock(&trace_types_lock
);
5765 if (cpu_id
!= RING_BUFFER_ALL_CPUS
) {
5766 /* make sure, this cpu is enabled in the mask */
5767 if (!cpumask_test_cpu(cpu_id
, tracing_buffer_mask
)) {
5773 ret
= __tracing_resize_ring_buffer(tr
, size
, cpu_id
);
5778 mutex_unlock(&trace_types_lock
);
5785 * tracing_update_buffers - used by tracing facility to expand ring buffers
5787 * To save on memory when the tracing is never used on a system with it
5788 * configured in. The ring buffers are set to a minimum size. But once
5789 * a user starts to use the tracing facility, then they need to grow
5790 * to their default size.
5792 * This function is to be called when a tracer is about to be used.
5794 int tracing_update_buffers(void)
5798 mutex_lock(&trace_types_lock
);
5799 if (!ring_buffer_expanded
)
5800 ret
= __tracing_resize_ring_buffer(&global_trace
, trace_buf_size
,
5801 RING_BUFFER_ALL_CPUS
);
5802 mutex_unlock(&trace_types_lock
);
5807 struct trace_option_dentry
;
5810 create_trace_option_files(struct trace_array
*tr
, struct tracer
*tracer
);
5813 * Used to clear out the tracer before deletion of an instance.
5814 * Must have trace_types_lock held.
5816 static void tracing_set_nop(struct trace_array
*tr
)
5818 if (tr
->current_trace
== &nop_trace
)
5821 tr
->current_trace
->enabled
--;
5823 if (tr
->current_trace
->reset
)
5824 tr
->current_trace
->reset(tr
);
5826 tr
->current_trace
= &nop_trace
;
5829 static void add_tracer_options(struct trace_array
*tr
, struct tracer
*t
)
5831 /* Only enable if the directory has been created already. */
5835 create_trace_option_files(tr
, t
);
5838 int tracing_set_tracer(struct trace_array
*tr
, const char *buf
)
5841 #ifdef CONFIG_TRACER_MAX_TRACE
5846 mutex_lock(&trace_types_lock
);
5848 if (!ring_buffer_expanded
) {
5849 ret
= __tracing_resize_ring_buffer(tr
, trace_buf_size
,
5850 RING_BUFFER_ALL_CPUS
);
5856 for (t
= trace_types
; t
; t
= t
->next
) {
5857 if (strcmp(t
->name
, buf
) == 0)
5864 if (t
== tr
->current_trace
)
5867 #ifdef CONFIG_TRACER_SNAPSHOT
5868 if (t
->use_max_tr
) {
5869 arch_spin_lock(&tr
->max_lock
);
5870 if (tr
->cond_snapshot
)
5872 arch_spin_unlock(&tr
->max_lock
);
5877 /* Some tracers won't work on kernel command line */
5878 if (system_state
< SYSTEM_RUNNING
&& t
->noboot
) {
5879 pr_warn("Tracer '%s' is not allowed on command line, ignored\n",
5884 /* Some tracers are only allowed for the top level buffer */
5885 if (!trace_ok_for_array(t
, tr
)) {
5890 /* If trace pipe files are being read, we can't change the tracer */
5891 if (tr
->current_trace
->ref
) {
5896 trace_branch_disable();
5898 tr
->current_trace
->enabled
--;
5900 if (tr
->current_trace
->reset
)
5901 tr
->current_trace
->reset(tr
);
5903 /* Current trace needs to be nop_trace before synchronize_rcu */
5904 tr
->current_trace
= &nop_trace
;
5906 #ifdef CONFIG_TRACER_MAX_TRACE
5907 had_max_tr
= tr
->allocated_snapshot
;
5909 if (had_max_tr
&& !t
->use_max_tr
) {
5911 * We need to make sure that the update_max_tr sees that
5912 * current_trace changed to nop_trace to keep it from
5913 * swapping the buffers after we resize it.
5914 * The update_max_tr is called from interrupts disabled
5915 * so a synchronized_sched() is sufficient.
5922 #ifdef CONFIG_TRACER_MAX_TRACE
5923 if (t
->use_max_tr
&& !had_max_tr
) {
5924 ret
= tracing_alloc_snapshot_instance(tr
);
5931 ret
= tracer_init(t
, tr
);
5936 tr
->current_trace
= t
;
5937 tr
->current_trace
->enabled
++;
5938 trace_branch_enable(tr
);
5940 mutex_unlock(&trace_types_lock
);
5946 tracing_set_trace_write(struct file
*filp
, const char __user
*ubuf
,
5947 size_t cnt
, loff_t
*ppos
)
5949 struct trace_array
*tr
= filp
->private_data
;
5950 char buf
[MAX_TRACER_SIZE
+1];
5957 if (cnt
> MAX_TRACER_SIZE
)
5958 cnt
= MAX_TRACER_SIZE
;
5960 if (copy_from_user(buf
, ubuf
, cnt
))
5965 /* strip ending whitespace. */
5966 for (i
= cnt
- 1; i
> 0 && isspace(buf
[i
]); i
--)
5969 err
= tracing_set_tracer(tr
, buf
);
5979 tracing_nsecs_read(unsigned long *ptr
, char __user
*ubuf
,
5980 size_t cnt
, loff_t
*ppos
)
5985 r
= snprintf(buf
, sizeof(buf
), "%ld\n",
5986 *ptr
== (unsigned long)-1 ? -1 : nsecs_to_usecs(*ptr
));
5987 if (r
> sizeof(buf
))
5989 return simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, r
);
5993 tracing_nsecs_write(unsigned long *ptr
, const char __user
*ubuf
,
5994 size_t cnt
, loff_t
*ppos
)
5999 ret
= kstrtoul_from_user(ubuf
, cnt
, 10, &val
);
6009 tracing_thresh_read(struct file
*filp
, char __user
*ubuf
,
6010 size_t cnt
, loff_t
*ppos
)
6012 return tracing_nsecs_read(&tracing_thresh
, ubuf
, cnt
, ppos
);
6016 tracing_thresh_write(struct file
*filp
, const char __user
*ubuf
,
6017 size_t cnt
, loff_t
*ppos
)
6019 struct trace_array
*tr
= filp
->private_data
;
6022 mutex_lock(&trace_types_lock
);
6023 ret
= tracing_nsecs_write(&tracing_thresh
, ubuf
, cnt
, ppos
);
6027 if (tr
->current_trace
->update_thresh
) {
6028 ret
= tr
->current_trace
->update_thresh(tr
);
6035 mutex_unlock(&trace_types_lock
);
6040 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)
6043 tracing_max_lat_read(struct file
*filp
, char __user
*ubuf
,
6044 size_t cnt
, loff_t
*ppos
)
6046 return tracing_nsecs_read(filp
->private_data
, ubuf
, cnt
, ppos
);
6050 tracing_max_lat_write(struct file
*filp
, const char __user
*ubuf
,
6051 size_t cnt
, loff_t
*ppos
)
6053 return tracing_nsecs_write(filp
->private_data
, ubuf
, cnt
, ppos
);
6058 static int tracing_open_pipe(struct inode
*inode
, struct file
*filp
)
6060 struct trace_array
*tr
= inode
->i_private
;
6061 struct trace_iterator
*iter
;
6064 ret
= tracing_check_open_get_tr(tr
);
6068 mutex_lock(&trace_types_lock
);
6070 /* create a buffer to store the information to pass to userspace */
6071 iter
= kzalloc(sizeof(*iter
), GFP_KERNEL
);
6074 __trace_array_put(tr
);
6078 trace_seq_init(&iter
->seq
);
6079 iter
->trace
= tr
->current_trace
;
6081 if (!alloc_cpumask_var(&iter
->started
, GFP_KERNEL
)) {
6086 /* trace pipe does not show start of buffer */
6087 cpumask_setall(iter
->started
);
6089 if (tr
->trace_flags
& TRACE_ITER_LATENCY_FMT
)
6090 iter
->iter_flags
|= TRACE_FILE_LAT_FMT
;
6092 /* Output in nanoseconds only if we are using a clock in nanoseconds. */
6093 if (trace_clocks
[tr
->clock_id
].in_ns
)
6094 iter
->iter_flags
|= TRACE_FILE_TIME_IN_NS
;
6097 iter
->array_buffer
= &tr
->array_buffer
;
6098 iter
->cpu_file
= tracing_get_cpu(inode
);
6099 mutex_init(&iter
->mutex
);
6100 filp
->private_data
= iter
;
6102 if (iter
->trace
->pipe_open
)
6103 iter
->trace
->pipe_open(iter
);
6105 nonseekable_open(inode
, filp
);
6107 tr
->current_trace
->ref
++;
6109 mutex_unlock(&trace_types_lock
);
6114 __trace_array_put(tr
);
6115 mutex_unlock(&trace_types_lock
);
6119 static int tracing_release_pipe(struct inode
*inode
, struct file
*file
)
6121 struct trace_iterator
*iter
= file
->private_data
;
6122 struct trace_array
*tr
= inode
->i_private
;
6124 mutex_lock(&trace_types_lock
);
6126 tr
->current_trace
->ref
--;
6128 if (iter
->trace
->pipe_close
)
6129 iter
->trace
->pipe_close(iter
);
6131 mutex_unlock(&trace_types_lock
);
6133 free_cpumask_var(iter
->started
);
6134 mutex_destroy(&iter
->mutex
);
6137 trace_array_put(tr
);
6143 trace_poll(struct trace_iterator
*iter
, struct file
*filp
, poll_table
*poll_table
)
6145 struct trace_array
*tr
= iter
->tr
;
6147 /* Iterators are static, they should be filled or empty */
6148 if (trace_buffer_iter(iter
, iter
->cpu_file
))
6149 return EPOLLIN
| EPOLLRDNORM
;
6151 if (tr
->trace_flags
& TRACE_ITER_BLOCK
)
6153 * Always select as readable when in blocking mode
6155 return EPOLLIN
| EPOLLRDNORM
;
6157 return ring_buffer_poll_wait(iter
->array_buffer
->buffer
, iter
->cpu_file
,
6162 tracing_poll_pipe(struct file
*filp
, poll_table
*poll_table
)
6164 struct trace_iterator
*iter
= filp
->private_data
;
6166 return trace_poll(iter
, filp
, poll_table
);
6169 /* Must be called with iter->mutex held. */
6170 static int tracing_wait_pipe(struct file
*filp
)
6172 struct trace_iterator
*iter
= filp
->private_data
;
6175 while (trace_empty(iter
)) {
6177 if ((filp
->f_flags
& O_NONBLOCK
)) {
6182 * We block until we read something and tracing is disabled.
6183 * We still block if tracing is disabled, but we have never
6184 * read anything. This allows a user to cat this file, and
6185 * then enable tracing. But after we have read something,
6186 * we give an EOF when tracing is again disabled.
6188 * iter->pos will be 0 if we haven't read anything.
6190 if (!tracer_tracing_is_on(iter
->tr
) && iter
->pos
)
6193 mutex_unlock(&iter
->mutex
);
6195 ret
= wait_on_pipe(iter
, 0);
6197 mutex_lock(&iter
->mutex
);
6210 tracing_read_pipe(struct file
*filp
, char __user
*ubuf
,
6211 size_t cnt
, loff_t
*ppos
)
6213 struct trace_iterator
*iter
= filp
->private_data
;
6217 * Avoid more than one consumer on a single file descriptor
6218 * This is just a matter of traces coherency, the ring buffer itself
6221 mutex_lock(&iter
->mutex
);
6223 /* return any leftover data */
6224 sret
= trace_seq_to_user(&iter
->seq
, ubuf
, cnt
);
6228 trace_seq_init(&iter
->seq
);
6230 if (iter
->trace
->read
) {
6231 sret
= iter
->trace
->read(iter
, filp
, ubuf
, cnt
, ppos
);
6237 sret
= tracing_wait_pipe(filp
);
6241 /* stop when tracing is finished */
6242 if (trace_empty(iter
)) {
6247 if (cnt
>= PAGE_SIZE
)
6248 cnt
= PAGE_SIZE
- 1;
6250 /* reset all but tr, trace, and overruns */
6251 memset(&iter
->seq
, 0,
6252 sizeof(struct trace_iterator
) -
6253 offsetof(struct trace_iterator
, seq
));
6254 cpumask_clear(iter
->started
);
6255 trace_seq_init(&iter
->seq
);
6258 trace_event_read_lock();
6259 trace_access_lock(iter
->cpu_file
);
6260 while (trace_find_next_entry_inc(iter
) != NULL
) {
6261 enum print_line_t ret
;
6262 int save_len
= iter
->seq
.seq
.len
;
6264 ret
= print_trace_line(iter
);
6265 if (ret
== TRACE_TYPE_PARTIAL_LINE
) {
6266 /* don't print partial lines */
6267 iter
->seq
.seq
.len
= save_len
;
6270 if (ret
!= TRACE_TYPE_NO_CONSUME
)
6271 trace_consume(iter
);
6273 if (trace_seq_used(&iter
->seq
) >= cnt
)
6277 * Setting the full flag means we reached the trace_seq buffer
6278 * size and we should leave by partial output condition above.
6279 * One of the trace_seq_* functions is not used properly.
6281 WARN_ONCE(iter
->seq
.full
, "full flag set for trace type %d",
6284 trace_access_unlock(iter
->cpu_file
);
6285 trace_event_read_unlock();
6287 /* Now copy what we have to the user */
6288 sret
= trace_seq_to_user(&iter
->seq
, ubuf
, cnt
);
6289 if (iter
->seq
.seq
.readpos
>= trace_seq_used(&iter
->seq
))
6290 trace_seq_init(&iter
->seq
);
6293 * If there was nothing to send to user, in spite of consuming trace
6294 * entries, go back to wait for more entries.
6300 mutex_unlock(&iter
->mutex
);
6305 static void tracing_spd_release_pipe(struct splice_pipe_desc
*spd
,
6308 __free_page(spd
->pages
[idx
]);
6312 tracing_fill_pipe_page(size_t rem
, struct trace_iterator
*iter
)
6318 /* Seq buffer is page-sized, exactly what we need. */
6320 save_len
= iter
->seq
.seq
.len
;
6321 ret
= print_trace_line(iter
);
6323 if (trace_seq_has_overflowed(&iter
->seq
)) {
6324 iter
->seq
.seq
.len
= save_len
;
6329 * This should not be hit, because it should only
6330 * be set if the iter->seq overflowed. But check it
6331 * anyway to be safe.
6333 if (ret
== TRACE_TYPE_PARTIAL_LINE
) {
6334 iter
->seq
.seq
.len
= save_len
;
6338 count
= trace_seq_used(&iter
->seq
) - save_len
;
6341 iter
->seq
.seq
.len
= save_len
;
6345 if (ret
!= TRACE_TYPE_NO_CONSUME
)
6346 trace_consume(iter
);
6348 if (!trace_find_next_entry_inc(iter
)) {
6358 static ssize_t
tracing_splice_read_pipe(struct file
*filp
,
6360 struct pipe_inode_info
*pipe
,
6364 struct page
*pages_def
[PIPE_DEF_BUFFERS
];
6365 struct partial_page partial_def
[PIPE_DEF_BUFFERS
];
6366 struct trace_iterator
*iter
= filp
->private_data
;
6367 struct splice_pipe_desc spd
= {
6369 .partial
= partial_def
,
6370 .nr_pages
= 0, /* This gets updated below. */
6371 .nr_pages_max
= PIPE_DEF_BUFFERS
,
6372 .ops
= &default_pipe_buf_ops
,
6373 .spd_release
= tracing_spd_release_pipe
,
6379 if (splice_grow_spd(pipe
, &spd
))
6382 mutex_lock(&iter
->mutex
);
6384 if (iter
->trace
->splice_read
) {
6385 ret
= iter
->trace
->splice_read(iter
, filp
,
6386 ppos
, pipe
, len
, flags
);
6391 ret
= tracing_wait_pipe(filp
);
6395 if (!iter
->ent
&& !trace_find_next_entry_inc(iter
)) {
6400 trace_event_read_lock();
6401 trace_access_lock(iter
->cpu_file
);
6403 /* Fill as many pages as possible. */
6404 for (i
= 0, rem
= len
; i
< spd
.nr_pages_max
&& rem
; i
++) {
6405 spd
.pages
[i
] = alloc_page(GFP_KERNEL
);
6409 rem
= tracing_fill_pipe_page(rem
, iter
);
6411 /* Copy the data into the page, so we can start over. */
6412 ret
= trace_seq_to_buffer(&iter
->seq
,
6413 page_address(spd
.pages
[i
]),
6414 trace_seq_used(&iter
->seq
));
6416 __free_page(spd
.pages
[i
]);
6419 spd
.partial
[i
].offset
= 0;
6420 spd
.partial
[i
].len
= trace_seq_used(&iter
->seq
);
6422 trace_seq_init(&iter
->seq
);
6425 trace_access_unlock(iter
->cpu_file
);
6426 trace_event_read_unlock();
6427 mutex_unlock(&iter
->mutex
);
6432 ret
= splice_to_pipe(pipe
, &spd
);
6436 splice_shrink_spd(&spd
);
6440 mutex_unlock(&iter
->mutex
);
6445 tracing_entries_read(struct file
*filp
, char __user
*ubuf
,
6446 size_t cnt
, loff_t
*ppos
)
6448 struct inode
*inode
= file_inode(filp
);
6449 struct trace_array
*tr
= inode
->i_private
;
6450 int cpu
= tracing_get_cpu(inode
);
6455 mutex_lock(&trace_types_lock
);
6457 if (cpu
== RING_BUFFER_ALL_CPUS
) {
6458 int cpu
, buf_size_same
;
6463 /* check if all cpu sizes are same */
6464 for_each_tracing_cpu(cpu
) {
6465 /* fill in the size from first enabled cpu */
6467 size
= per_cpu_ptr(tr
->array_buffer
.data
, cpu
)->entries
;
6468 if (size
!= per_cpu_ptr(tr
->array_buffer
.data
, cpu
)->entries
) {
6474 if (buf_size_same
) {
6475 if (!ring_buffer_expanded
)
6476 r
= sprintf(buf
, "%lu (expanded: %lu)\n",
6478 trace_buf_size
>> 10);
6480 r
= sprintf(buf
, "%lu\n", size
>> 10);
6482 r
= sprintf(buf
, "X\n");
6484 r
= sprintf(buf
, "%lu\n", per_cpu_ptr(tr
->array_buffer
.data
, cpu
)->entries
>> 10);
6486 mutex_unlock(&trace_types_lock
);
6488 ret
= simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, r
);
6493 tracing_entries_write(struct file
*filp
, const char __user
*ubuf
,
6494 size_t cnt
, loff_t
*ppos
)
6496 struct inode
*inode
= file_inode(filp
);
6497 struct trace_array
*tr
= inode
->i_private
;
6501 ret
= kstrtoul_from_user(ubuf
, cnt
, 10, &val
);
6505 /* must have at least 1 entry */
6509 /* value is in KB */
6511 ret
= tracing_resize_ring_buffer(tr
, val
, tracing_get_cpu(inode
));
6521 tracing_total_entries_read(struct file
*filp
, char __user
*ubuf
,
6522 size_t cnt
, loff_t
*ppos
)
6524 struct trace_array
*tr
= filp
->private_data
;
6527 unsigned long size
= 0, expanded_size
= 0;
6529 mutex_lock(&trace_types_lock
);
6530 for_each_tracing_cpu(cpu
) {
6531 size
+= per_cpu_ptr(tr
->array_buffer
.data
, cpu
)->entries
>> 10;
6532 if (!ring_buffer_expanded
)
6533 expanded_size
+= trace_buf_size
>> 10;
6535 if (ring_buffer_expanded
)
6536 r
= sprintf(buf
, "%lu\n", size
);
6538 r
= sprintf(buf
, "%lu (expanded: %lu)\n", size
, expanded_size
);
6539 mutex_unlock(&trace_types_lock
);
6541 return simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, r
);
6545 tracing_free_buffer_write(struct file
*filp
, const char __user
*ubuf
,
6546 size_t cnt
, loff_t
*ppos
)
6549 * There is no need to read what the user has written, this function
6550 * is just to make sure that there is no error when "echo" is used
6559 tracing_free_buffer_release(struct inode
*inode
, struct file
*filp
)
6561 struct trace_array
*tr
= inode
->i_private
;
6563 /* disable tracing ? */
6564 if (tr
->trace_flags
& TRACE_ITER_STOP_ON_FREE
)
6565 tracer_tracing_off(tr
);
6566 /* resize the ring buffer to 0 */
6567 tracing_resize_ring_buffer(tr
, 0, RING_BUFFER_ALL_CPUS
);
6569 trace_array_put(tr
);
6575 tracing_mark_write(struct file
*filp
, const char __user
*ubuf
,
6576 size_t cnt
, loff_t
*fpos
)
6578 struct trace_array
*tr
= filp
->private_data
;
6579 struct ring_buffer_event
*event
;
6580 enum event_trigger_type tt
= ETT_NONE
;
6581 struct trace_buffer
*buffer
;
6582 struct print_entry
*entry
;
6583 unsigned long irq_flags
;
6588 /* Used in tracing_mark_raw_write() as well */
6589 #define FAULTED_STR "<faulted>"
6590 #define FAULTED_SIZE (sizeof(FAULTED_STR) - 1) /* '\0' is already accounted for */
6592 if (tracing_disabled
)
6595 if (!(tr
->trace_flags
& TRACE_ITER_MARKERS
))
6598 if (cnt
> TRACE_BUF_SIZE
)
6599 cnt
= TRACE_BUF_SIZE
;
6601 BUILD_BUG_ON(TRACE_BUF_SIZE
>= PAGE_SIZE
);
6603 local_save_flags(irq_flags
);
6604 size
= sizeof(*entry
) + cnt
+ 2; /* add '\0' and possible '\n' */
6606 /* If less than "<faulted>", then make sure we can still add that */
6607 if (cnt
< FAULTED_SIZE
)
6608 size
+= FAULTED_SIZE
- cnt
;
6610 buffer
= tr
->array_buffer
.buffer
;
6611 event
= __trace_buffer_lock_reserve(buffer
, TRACE_PRINT
, size
,
6612 irq_flags
, preempt_count());
6613 if (unlikely(!event
))
6614 /* Ring buffer disabled, return as if not open for write */
6617 entry
= ring_buffer_event_data(event
);
6618 entry
->ip
= _THIS_IP_
;
6620 len
= __copy_from_user_inatomic(&entry
->buf
, ubuf
, cnt
);
6622 memcpy(&entry
->buf
, FAULTED_STR
, FAULTED_SIZE
);
6629 if (tr
->trace_marker_file
&& !list_empty(&tr
->trace_marker_file
->triggers
)) {
6630 /* do not add \n before testing triggers, but add \0 */
6631 entry
->buf
[cnt
] = '\0';
6632 tt
= event_triggers_call(tr
->trace_marker_file
, entry
, event
);
6635 if (entry
->buf
[cnt
- 1] != '\n') {
6636 entry
->buf
[cnt
] = '\n';
6637 entry
->buf
[cnt
+ 1] = '\0';
6639 entry
->buf
[cnt
] = '\0';
6641 __buffer_unlock_commit(buffer
, event
);
6644 event_triggers_post_call(tr
->trace_marker_file
, tt
);
6652 /* Limit it for now to 3K (including tag) */
6653 #define RAW_DATA_MAX_SIZE (1024*3)
6656 tracing_mark_raw_write(struct file
*filp
, const char __user
*ubuf
,
6657 size_t cnt
, loff_t
*fpos
)
6659 struct trace_array
*tr
= filp
->private_data
;
6660 struct ring_buffer_event
*event
;
6661 struct trace_buffer
*buffer
;
6662 struct raw_data_entry
*entry
;
6663 unsigned long irq_flags
;
6668 #define FAULT_SIZE_ID (FAULTED_SIZE + sizeof(int))
6670 if (tracing_disabled
)
6673 if (!(tr
->trace_flags
& TRACE_ITER_MARKERS
))
6676 /* The marker must at least have a tag id */
6677 if (cnt
< sizeof(unsigned int) || cnt
> RAW_DATA_MAX_SIZE
)
6680 if (cnt
> TRACE_BUF_SIZE
)
6681 cnt
= TRACE_BUF_SIZE
;
6683 BUILD_BUG_ON(TRACE_BUF_SIZE
>= PAGE_SIZE
);
6685 local_save_flags(irq_flags
);
6686 size
= sizeof(*entry
) + cnt
;
6687 if (cnt
< FAULT_SIZE_ID
)
6688 size
+= FAULT_SIZE_ID
- cnt
;
6690 buffer
= tr
->array_buffer
.buffer
;
6691 event
= __trace_buffer_lock_reserve(buffer
, TRACE_RAW_DATA
, size
,
6692 irq_flags
, preempt_count());
6694 /* Ring buffer disabled, return as if not open for write */
6697 entry
= ring_buffer_event_data(event
);
6699 len
= __copy_from_user_inatomic(&entry
->id
, ubuf
, cnt
);
6702 memcpy(&entry
->buf
, FAULTED_STR
, FAULTED_SIZE
);
6707 __buffer_unlock_commit(buffer
, event
);
6715 static int tracing_clock_show(struct seq_file
*m
, void *v
)
6717 struct trace_array
*tr
= m
->private;
6720 for (i
= 0; i
< ARRAY_SIZE(trace_clocks
); i
++)
6722 "%s%s%s%s", i
? " " : "",
6723 i
== tr
->clock_id
? "[" : "", trace_clocks
[i
].name
,
6724 i
== tr
->clock_id
? "]" : "");
6730 int tracing_set_clock(struct trace_array
*tr
, const char *clockstr
)
6734 for (i
= 0; i
< ARRAY_SIZE(trace_clocks
); i
++) {
6735 if (strcmp(trace_clocks
[i
].name
, clockstr
) == 0)
6738 if (i
== ARRAY_SIZE(trace_clocks
))
6741 mutex_lock(&trace_types_lock
);
6745 ring_buffer_set_clock(tr
->array_buffer
.buffer
, trace_clocks
[i
].func
);
6748 * New clock may not be consistent with the previous clock.
6749 * Reset the buffer so that it doesn't have incomparable timestamps.
6751 tracing_reset_online_cpus(&tr
->array_buffer
);
6753 #ifdef CONFIG_TRACER_MAX_TRACE
6754 if (tr
->max_buffer
.buffer
)
6755 ring_buffer_set_clock(tr
->max_buffer
.buffer
, trace_clocks
[i
].func
);
6756 tracing_reset_online_cpus(&tr
->max_buffer
);
6759 mutex_unlock(&trace_types_lock
);
6764 static ssize_t
tracing_clock_write(struct file
*filp
, const char __user
*ubuf
,
6765 size_t cnt
, loff_t
*fpos
)
6767 struct seq_file
*m
= filp
->private_data
;
6768 struct trace_array
*tr
= m
->private;
6770 const char *clockstr
;
6773 if (cnt
>= sizeof(buf
))
6776 if (copy_from_user(buf
, ubuf
, cnt
))
6781 clockstr
= strstrip(buf
);
6783 ret
= tracing_set_clock(tr
, clockstr
);
6792 static int tracing_clock_open(struct inode
*inode
, struct file
*file
)
6794 struct trace_array
*tr
= inode
->i_private
;
6797 ret
= tracing_check_open_get_tr(tr
);
6801 ret
= single_open(file
, tracing_clock_show
, inode
->i_private
);
6803 trace_array_put(tr
);
6808 static int tracing_time_stamp_mode_show(struct seq_file
*m
, void *v
)
6810 struct trace_array
*tr
= m
->private;
6812 mutex_lock(&trace_types_lock
);
6814 if (ring_buffer_time_stamp_abs(tr
->array_buffer
.buffer
))
6815 seq_puts(m
, "delta [absolute]\n");
6817 seq_puts(m
, "[delta] absolute\n");
6819 mutex_unlock(&trace_types_lock
);
6824 static int tracing_time_stamp_mode_open(struct inode
*inode
, struct file
*file
)
6826 struct trace_array
*tr
= inode
->i_private
;
6829 ret
= tracing_check_open_get_tr(tr
);
6833 ret
= single_open(file
, tracing_time_stamp_mode_show
, inode
->i_private
);
6835 trace_array_put(tr
);
6840 int tracing_set_time_stamp_abs(struct trace_array
*tr
, bool abs
)
6844 mutex_lock(&trace_types_lock
);
6846 if (abs
&& tr
->time_stamp_abs_ref
++)
6850 if (WARN_ON_ONCE(!tr
->time_stamp_abs_ref
)) {
6855 if (--tr
->time_stamp_abs_ref
)
6859 ring_buffer_set_time_stamp_abs(tr
->array_buffer
.buffer
, abs
);
6861 #ifdef CONFIG_TRACER_MAX_TRACE
6862 if (tr
->max_buffer
.buffer
)
6863 ring_buffer_set_time_stamp_abs(tr
->max_buffer
.buffer
, abs
);
6866 mutex_unlock(&trace_types_lock
);
6871 struct ftrace_buffer_info
{
6872 struct trace_iterator iter
;
6874 unsigned int spare_cpu
;
6878 #ifdef CONFIG_TRACER_SNAPSHOT
6879 static int tracing_snapshot_open(struct inode
*inode
, struct file
*file
)
6881 struct trace_array
*tr
= inode
->i_private
;
6882 struct trace_iterator
*iter
;
6886 ret
= tracing_check_open_get_tr(tr
);
6890 if (file
->f_mode
& FMODE_READ
) {
6891 iter
= __tracing_open(inode
, file
, true);
6893 ret
= PTR_ERR(iter
);
6895 /* Writes still need the seq_file to hold the private data */
6897 m
= kzalloc(sizeof(*m
), GFP_KERNEL
);
6900 iter
= kzalloc(sizeof(*iter
), GFP_KERNEL
);
6908 iter
->array_buffer
= &tr
->max_buffer
;
6909 iter
->cpu_file
= tracing_get_cpu(inode
);
6911 file
->private_data
= m
;
6915 trace_array_put(tr
);
6921 tracing_snapshot_write(struct file
*filp
, const char __user
*ubuf
, size_t cnt
,
6924 struct seq_file
*m
= filp
->private_data
;
6925 struct trace_iterator
*iter
= m
->private;
6926 struct trace_array
*tr
= iter
->tr
;
6930 ret
= tracing_update_buffers();
6934 ret
= kstrtoul_from_user(ubuf
, cnt
, 10, &val
);
6938 mutex_lock(&trace_types_lock
);
6940 if (tr
->current_trace
->use_max_tr
) {
6945 arch_spin_lock(&tr
->max_lock
);
6946 if (tr
->cond_snapshot
)
6948 arch_spin_unlock(&tr
->max_lock
);
6954 if (iter
->cpu_file
!= RING_BUFFER_ALL_CPUS
) {
6958 if (tr
->allocated_snapshot
)
6962 /* Only allow per-cpu swap if the ring buffer supports it */
6963 #ifndef CONFIG_RING_BUFFER_ALLOW_SWAP
6964 if (iter
->cpu_file
!= RING_BUFFER_ALL_CPUS
) {
6969 if (tr
->allocated_snapshot
)
6970 ret
= resize_buffer_duplicate_size(&tr
->max_buffer
,
6971 &tr
->array_buffer
, iter
->cpu_file
);
6973 ret
= tracing_alloc_snapshot_instance(tr
);
6976 local_irq_disable();
6977 /* Now, we're going to swap */
6978 if (iter
->cpu_file
== RING_BUFFER_ALL_CPUS
)
6979 update_max_tr(tr
, current
, smp_processor_id(), NULL
);
6981 update_max_tr_single(tr
, current
, iter
->cpu_file
);
6985 if (tr
->allocated_snapshot
) {
6986 if (iter
->cpu_file
== RING_BUFFER_ALL_CPUS
)
6987 tracing_reset_online_cpus(&tr
->max_buffer
);
6989 tracing_reset_cpu(&tr
->max_buffer
, iter
->cpu_file
);
6999 mutex_unlock(&trace_types_lock
);
7003 static int tracing_snapshot_release(struct inode
*inode
, struct file
*file
)
7005 struct seq_file
*m
= file
->private_data
;
7008 ret
= tracing_release(inode
, file
);
7010 if (file
->f_mode
& FMODE_READ
)
7013 /* If write only, the seq_file is just a stub */
7021 static int tracing_buffers_open(struct inode
*inode
, struct file
*filp
);
7022 static ssize_t
tracing_buffers_read(struct file
*filp
, char __user
*ubuf
,
7023 size_t count
, loff_t
*ppos
);
7024 static int tracing_buffers_release(struct inode
*inode
, struct file
*file
);
7025 static ssize_t
tracing_buffers_splice_read(struct file
*file
, loff_t
*ppos
,
7026 struct pipe_inode_info
*pipe
, size_t len
, unsigned int flags
);
7028 static int snapshot_raw_open(struct inode
*inode
, struct file
*filp
)
7030 struct ftrace_buffer_info
*info
;
7033 /* The following checks for tracefs lockdown */
7034 ret
= tracing_buffers_open(inode
, filp
);
7038 info
= filp
->private_data
;
7040 if (info
->iter
.trace
->use_max_tr
) {
7041 tracing_buffers_release(inode
, filp
);
7045 info
->iter
.snapshot
= true;
7046 info
->iter
.array_buffer
= &info
->iter
.tr
->max_buffer
;
7051 #endif /* CONFIG_TRACER_SNAPSHOT */
7054 static const struct file_operations tracing_thresh_fops
= {
7055 .open
= tracing_open_generic
,
7056 .read
= tracing_thresh_read
,
7057 .write
= tracing_thresh_write
,
7058 .llseek
= generic_file_llseek
,
7061 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)
7062 static const struct file_operations tracing_max_lat_fops
= {
7063 .open
= tracing_open_generic
,
7064 .read
= tracing_max_lat_read
,
7065 .write
= tracing_max_lat_write
,
7066 .llseek
= generic_file_llseek
,
7070 static const struct file_operations set_tracer_fops
= {
7071 .open
= tracing_open_generic
,
7072 .read
= tracing_set_trace_read
,
7073 .write
= tracing_set_trace_write
,
7074 .llseek
= generic_file_llseek
,
7077 static const struct file_operations tracing_pipe_fops
= {
7078 .open
= tracing_open_pipe
,
7079 .poll
= tracing_poll_pipe
,
7080 .read
= tracing_read_pipe
,
7081 .splice_read
= tracing_splice_read_pipe
,
7082 .release
= tracing_release_pipe
,
7083 .llseek
= no_llseek
,
7086 static const struct file_operations tracing_entries_fops
= {
7087 .open
= tracing_open_generic_tr
,
7088 .read
= tracing_entries_read
,
7089 .write
= tracing_entries_write
,
7090 .llseek
= generic_file_llseek
,
7091 .release
= tracing_release_generic_tr
,
7094 static const struct file_operations tracing_total_entries_fops
= {
7095 .open
= tracing_open_generic_tr
,
7096 .read
= tracing_total_entries_read
,
7097 .llseek
= generic_file_llseek
,
7098 .release
= tracing_release_generic_tr
,
7101 static const struct file_operations tracing_free_buffer_fops
= {
7102 .open
= tracing_open_generic_tr
,
7103 .write
= tracing_free_buffer_write
,
7104 .release
= tracing_free_buffer_release
,
7107 static const struct file_operations tracing_mark_fops
= {
7108 .open
= tracing_open_generic_tr
,
7109 .write
= tracing_mark_write
,
7110 .llseek
= generic_file_llseek
,
7111 .release
= tracing_release_generic_tr
,
7114 static const struct file_operations tracing_mark_raw_fops
= {
7115 .open
= tracing_open_generic_tr
,
7116 .write
= tracing_mark_raw_write
,
7117 .llseek
= generic_file_llseek
,
7118 .release
= tracing_release_generic_tr
,
7121 static const struct file_operations trace_clock_fops
= {
7122 .open
= tracing_clock_open
,
7124 .llseek
= seq_lseek
,
7125 .release
= tracing_single_release_tr
,
7126 .write
= tracing_clock_write
,
7129 static const struct file_operations trace_time_stamp_mode_fops
= {
7130 .open
= tracing_time_stamp_mode_open
,
7132 .llseek
= seq_lseek
,
7133 .release
= tracing_single_release_tr
,
7136 #ifdef CONFIG_TRACER_SNAPSHOT
7137 static const struct file_operations snapshot_fops
= {
7138 .open
= tracing_snapshot_open
,
7140 .write
= tracing_snapshot_write
,
7141 .llseek
= tracing_lseek
,
7142 .release
= tracing_snapshot_release
,
7145 static const struct file_operations snapshot_raw_fops
= {
7146 .open
= snapshot_raw_open
,
7147 .read
= tracing_buffers_read
,
7148 .release
= tracing_buffers_release
,
7149 .splice_read
= tracing_buffers_splice_read
,
7150 .llseek
= no_llseek
,
7153 #endif /* CONFIG_TRACER_SNAPSHOT */
7155 #define TRACING_LOG_ERRS_MAX 8
7156 #define TRACING_LOG_LOC_MAX 128
7158 #define CMD_PREFIX " Command: "
7161 const char **errs
; /* ptr to loc-specific array of err strings */
7162 u8 type
; /* index into errs -> specific err string */
7163 u8 pos
; /* MAX_FILTER_STR_VAL = 256 */
7167 struct tracing_log_err
{
7168 struct list_head list
;
7169 struct err_info info
;
7170 char loc
[TRACING_LOG_LOC_MAX
]; /* err location */
7171 char cmd
[MAX_FILTER_STR_VAL
]; /* what caused err */
7174 static DEFINE_MUTEX(tracing_err_log_lock
);
7176 static struct tracing_log_err
*get_tracing_log_err(struct trace_array
*tr
)
7178 struct tracing_log_err
*err
;
7180 if (tr
->n_err_log_entries
< TRACING_LOG_ERRS_MAX
) {
7181 err
= kzalloc(sizeof(*err
), GFP_KERNEL
);
7183 err
= ERR_PTR(-ENOMEM
);
7184 tr
->n_err_log_entries
++;
7189 err
= list_first_entry(&tr
->err_log
, struct tracing_log_err
, list
);
7190 list_del(&err
->list
);
7196 * err_pos - find the position of a string within a command for error careting
7197 * @cmd: The tracing command that caused the error
7198 * @str: The string to position the caret at within @cmd
7200 * Finds the position of the first occurence of @str within @cmd. The
7201 * return value can be passed to tracing_log_err() for caret placement
7204 * Returns the index within @cmd of the first occurence of @str or 0
7205 * if @str was not found.
7207 unsigned int err_pos(char *cmd
, const char *str
)
7211 if (WARN_ON(!strlen(cmd
)))
7214 found
= strstr(cmd
, str
);
7222 * tracing_log_err - write an error to the tracing error log
7223 * @tr: The associated trace array for the error (NULL for top level array)
7224 * @loc: A string describing where the error occurred
7225 * @cmd: The tracing command that caused the error
7226 * @errs: The array of loc-specific static error strings
7227 * @type: The index into errs[], which produces the specific static err string
7228 * @pos: The position the caret should be placed in the cmd
7230 * Writes an error into tracing/error_log of the form:
7232 * <loc>: error: <text>
7236 * tracing/error_log is a small log file containing the last
7237 * TRACING_LOG_ERRS_MAX errors (8). Memory for errors isn't allocated
7238 * unless there has been a tracing error, and the error log can be
7239 * cleared and have its memory freed by writing the empty string in
7240 * truncation mode to it i.e. echo > tracing/error_log.
7242 * NOTE: the @errs array along with the @type param are used to
7243 * produce a static error string - this string is not copied and saved
7244 * when the error is logged - only a pointer to it is saved. See
7245 * existing callers for examples of how static strings are typically
7246 * defined for use with tracing_log_err().
7248 void tracing_log_err(struct trace_array
*tr
,
7249 const char *loc
, const char *cmd
,
7250 const char **errs
, u8 type
, u8 pos
)
7252 struct tracing_log_err
*err
;
7257 mutex_lock(&tracing_err_log_lock
);
7258 err
= get_tracing_log_err(tr
);
7259 if (PTR_ERR(err
) == -ENOMEM
) {
7260 mutex_unlock(&tracing_err_log_lock
);
7264 snprintf(err
->loc
, TRACING_LOG_LOC_MAX
, "%s: error: ", loc
);
7265 snprintf(err
->cmd
, MAX_FILTER_STR_VAL
,"\n" CMD_PREFIX
"%s\n", cmd
);
7267 err
->info
.errs
= errs
;
7268 err
->info
.type
= type
;
7269 err
->info
.pos
= pos
;
7270 err
->info
.ts
= local_clock();
7272 list_add_tail(&err
->list
, &tr
->err_log
);
7273 mutex_unlock(&tracing_err_log_lock
);
7276 static void clear_tracing_err_log(struct trace_array
*tr
)
7278 struct tracing_log_err
*err
, *next
;
7280 mutex_lock(&tracing_err_log_lock
);
7281 list_for_each_entry_safe(err
, next
, &tr
->err_log
, list
) {
7282 list_del(&err
->list
);
7286 tr
->n_err_log_entries
= 0;
7287 mutex_unlock(&tracing_err_log_lock
);
7290 static void *tracing_err_log_seq_start(struct seq_file
*m
, loff_t
*pos
)
7292 struct trace_array
*tr
= m
->private;
7294 mutex_lock(&tracing_err_log_lock
);
7296 return seq_list_start(&tr
->err_log
, *pos
);
7299 static void *tracing_err_log_seq_next(struct seq_file
*m
, void *v
, loff_t
*pos
)
7301 struct trace_array
*tr
= m
->private;
7303 return seq_list_next(v
, &tr
->err_log
, pos
);
7306 static void tracing_err_log_seq_stop(struct seq_file
*m
, void *v
)
7308 mutex_unlock(&tracing_err_log_lock
);
7311 static void tracing_err_log_show_pos(struct seq_file
*m
, u8 pos
)
7315 for (i
= 0; i
< sizeof(CMD_PREFIX
) - 1; i
++)
7317 for (i
= 0; i
< pos
; i
++)
7322 static int tracing_err_log_seq_show(struct seq_file
*m
, void *v
)
7324 struct tracing_log_err
*err
= v
;
7327 const char *err_text
= err
->info
.errs
[err
->info
.type
];
7328 u64 sec
= err
->info
.ts
;
7331 nsec
= do_div(sec
, NSEC_PER_SEC
);
7332 seq_printf(m
, "[%5llu.%06u] %s%s", sec
, nsec
/ 1000,
7333 err
->loc
, err_text
);
7334 seq_printf(m
, "%s", err
->cmd
);
7335 tracing_err_log_show_pos(m
, err
->info
.pos
);
7341 static const struct seq_operations tracing_err_log_seq_ops
= {
7342 .start
= tracing_err_log_seq_start
,
7343 .next
= tracing_err_log_seq_next
,
7344 .stop
= tracing_err_log_seq_stop
,
7345 .show
= tracing_err_log_seq_show
7348 static int tracing_err_log_open(struct inode
*inode
, struct file
*file
)
7350 struct trace_array
*tr
= inode
->i_private
;
7353 ret
= tracing_check_open_get_tr(tr
);
7357 /* If this file was opened for write, then erase contents */
7358 if ((file
->f_mode
& FMODE_WRITE
) && (file
->f_flags
& O_TRUNC
))
7359 clear_tracing_err_log(tr
);
7361 if (file
->f_mode
& FMODE_READ
) {
7362 ret
= seq_open(file
, &tracing_err_log_seq_ops
);
7364 struct seq_file
*m
= file
->private_data
;
7367 trace_array_put(tr
);
7373 static ssize_t
tracing_err_log_write(struct file
*file
,
7374 const char __user
*buffer
,
7375 size_t count
, loff_t
*ppos
)
7380 static int tracing_err_log_release(struct inode
*inode
, struct file
*file
)
7382 struct trace_array
*tr
= inode
->i_private
;
7384 trace_array_put(tr
);
7386 if (file
->f_mode
& FMODE_READ
)
7387 seq_release(inode
, file
);
7392 static const struct file_operations tracing_err_log_fops
= {
7393 .open
= tracing_err_log_open
,
7394 .write
= tracing_err_log_write
,
7396 .llseek
= seq_lseek
,
7397 .release
= tracing_err_log_release
,
7400 static int tracing_buffers_open(struct inode
*inode
, struct file
*filp
)
7402 struct trace_array
*tr
= inode
->i_private
;
7403 struct ftrace_buffer_info
*info
;
7406 ret
= tracing_check_open_get_tr(tr
);
7410 info
= kzalloc(sizeof(*info
), GFP_KERNEL
);
7412 trace_array_put(tr
);
7416 mutex_lock(&trace_types_lock
);
7419 info
->iter
.cpu_file
= tracing_get_cpu(inode
);
7420 info
->iter
.trace
= tr
->current_trace
;
7421 info
->iter
.array_buffer
= &tr
->array_buffer
;
7423 /* Force reading ring buffer for first read */
7424 info
->read
= (unsigned int)-1;
7426 filp
->private_data
= info
;
7428 tr
->current_trace
->ref
++;
7430 mutex_unlock(&trace_types_lock
);
7432 ret
= nonseekable_open(inode
, filp
);
7434 trace_array_put(tr
);
7440 tracing_buffers_poll(struct file
*filp
, poll_table
*poll_table
)
7442 struct ftrace_buffer_info
*info
= filp
->private_data
;
7443 struct trace_iterator
*iter
= &info
->iter
;
7445 return trace_poll(iter
, filp
, poll_table
);
7449 tracing_buffers_read(struct file
*filp
, char __user
*ubuf
,
7450 size_t count
, loff_t
*ppos
)
7452 struct ftrace_buffer_info
*info
= filp
->private_data
;
7453 struct trace_iterator
*iter
= &info
->iter
;
7460 #ifdef CONFIG_TRACER_MAX_TRACE
7461 if (iter
->snapshot
&& iter
->tr
->current_trace
->use_max_tr
)
7466 info
->spare
= ring_buffer_alloc_read_page(iter
->array_buffer
->buffer
,
7468 if (IS_ERR(info
->spare
)) {
7469 ret
= PTR_ERR(info
->spare
);
7472 info
->spare_cpu
= iter
->cpu_file
;
7478 /* Do we have previous read data to read? */
7479 if (info
->read
< PAGE_SIZE
)
7483 trace_access_lock(iter
->cpu_file
);
7484 ret
= ring_buffer_read_page(iter
->array_buffer
->buffer
,
7488 trace_access_unlock(iter
->cpu_file
);
7491 if (trace_empty(iter
)) {
7492 if ((filp
->f_flags
& O_NONBLOCK
))
7495 ret
= wait_on_pipe(iter
, 0);
7506 size
= PAGE_SIZE
- info
->read
;
7510 ret
= copy_to_user(ubuf
, info
->spare
+ info
->read
, size
);
7522 static int tracing_buffers_release(struct inode
*inode
, struct file
*file
)
7524 struct ftrace_buffer_info
*info
= file
->private_data
;
7525 struct trace_iterator
*iter
= &info
->iter
;
7527 mutex_lock(&trace_types_lock
);
7529 iter
->tr
->current_trace
->ref
--;
7531 __trace_array_put(iter
->tr
);
7534 ring_buffer_free_read_page(iter
->array_buffer
->buffer
,
7535 info
->spare_cpu
, info
->spare
);
7538 mutex_unlock(&trace_types_lock
);
7544 struct trace_buffer
*buffer
;
7547 refcount_t refcount
;
7550 static void buffer_ref_release(struct buffer_ref
*ref
)
7552 if (!refcount_dec_and_test(&ref
->refcount
))
7554 ring_buffer_free_read_page(ref
->buffer
, ref
->cpu
, ref
->page
);
7558 static void buffer_pipe_buf_release(struct pipe_inode_info
*pipe
,
7559 struct pipe_buffer
*buf
)
7561 struct buffer_ref
*ref
= (struct buffer_ref
*)buf
->private;
7563 buffer_ref_release(ref
);
7567 static bool buffer_pipe_buf_get(struct pipe_inode_info
*pipe
,
7568 struct pipe_buffer
*buf
)
7570 struct buffer_ref
*ref
= (struct buffer_ref
*)buf
->private;
7572 if (refcount_read(&ref
->refcount
) > INT_MAX
/2)
7575 refcount_inc(&ref
->refcount
);
7579 /* Pipe buffer operations for a buffer. */
7580 static const struct pipe_buf_operations buffer_pipe_buf_ops
= {
7581 .release
= buffer_pipe_buf_release
,
7582 .get
= buffer_pipe_buf_get
,
7586 * Callback from splice_to_pipe(), if we need to release some pages
7587 * at the end of the spd in case we error'ed out in filling the pipe.
7589 static void buffer_spd_release(struct splice_pipe_desc
*spd
, unsigned int i
)
7591 struct buffer_ref
*ref
=
7592 (struct buffer_ref
*)spd
->partial
[i
].private;
7594 buffer_ref_release(ref
);
7595 spd
->partial
[i
].private = 0;
7599 tracing_buffers_splice_read(struct file
*file
, loff_t
*ppos
,
7600 struct pipe_inode_info
*pipe
, size_t len
,
7603 struct ftrace_buffer_info
*info
= file
->private_data
;
7604 struct trace_iterator
*iter
= &info
->iter
;
7605 struct partial_page partial_def
[PIPE_DEF_BUFFERS
];
7606 struct page
*pages_def
[PIPE_DEF_BUFFERS
];
7607 struct splice_pipe_desc spd
= {
7609 .partial
= partial_def
,
7610 .nr_pages_max
= PIPE_DEF_BUFFERS
,
7611 .ops
= &buffer_pipe_buf_ops
,
7612 .spd_release
= buffer_spd_release
,
7614 struct buffer_ref
*ref
;
7618 #ifdef CONFIG_TRACER_MAX_TRACE
7619 if (iter
->snapshot
&& iter
->tr
->current_trace
->use_max_tr
)
7623 if (*ppos
& (PAGE_SIZE
- 1))
7626 if (len
& (PAGE_SIZE
- 1)) {
7627 if (len
< PAGE_SIZE
)
7632 if (splice_grow_spd(pipe
, &spd
))
7636 trace_access_lock(iter
->cpu_file
);
7637 entries
= ring_buffer_entries_cpu(iter
->array_buffer
->buffer
, iter
->cpu_file
);
7639 for (i
= 0; i
< spd
.nr_pages_max
&& len
&& entries
; i
++, len
-= PAGE_SIZE
) {
7643 ref
= kzalloc(sizeof(*ref
), GFP_KERNEL
);
7649 refcount_set(&ref
->refcount
, 1);
7650 ref
->buffer
= iter
->array_buffer
->buffer
;
7651 ref
->page
= ring_buffer_alloc_read_page(ref
->buffer
, iter
->cpu_file
);
7652 if (IS_ERR(ref
->page
)) {
7653 ret
= PTR_ERR(ref
->page
);
7658 ref
->cpu
= iter
->cpu_file
;
7660 r
= ring_buffer_read_page(ref
->buffer
, &ref
->page
,
7661 len
, iter
->cpu_file
, 1);
7663 ring_buffer_free_read_page(ref
->buffer
, ref
->cpu
,
7669 page
= virt_to_page(ref
->page
);
7671 spd
.pages
[i
] = page
;
7672 spd
.partial
[i
].len
= PAGE_SIZE
;
7673 spd
.partial
[i
].offset
= 0;
7674 spd
.partial
[i
].private = (unsigned long)ref
;
7678 entries
= ring_buffer_entries_cpu(iter
->array_buffer
->buffer
, iter
->cpu_file
);
7681 trace_access_unlock(iter
->cpu_file
);
7684 /* did we read anything? */
7685 if (!spd
.nr_pages
) {
7690 if ((file
->f_flags
& O_NONBLOCK
) || (flags
& SPLICE_F_NONBLOCK
))
7693 ret
= wait_on_pipe(iter
, iter
->tr
->buffer_percent
);
7700 ret
= splice_to_pipe(pipe
, &spd
);
7702 splice_shrink_spd(&spd
);
7707 static const struct file_operations tracing_buffers_fops
= {
7708 .open
= tracing_buffers_open
,
7709 .read
= tracing_buffers_read
,
7710 .poll
= tracing_buffers_poll
,
7711 .release
= tracing_buffers_release
,
7712 .splice_read
= tracing_buffers_splice_read
,
7713 .llseek
= no_llseek
,
7717 tracing_stats_read(struct file
*filp
, char __user
*ubuf
,
7718 size_t count
, loff_t
*ppos
)
7720 struct inode
*inode
= file_inode(filp
);
7721 struct trace_array
*tr
= inode
->i_private
;
7722 struct array_buffer
*trace_buf
= &tr
->array_buffer
;
7723 int cpu
= tracing_get_cpu(inode
);
7724 struct trace_seq
*s
;
7726 unsigned long long t
;
7727 unsigned long usec_rem
;
7729 s
= kmalloc(sizeof(*s
), GFP_KERNEL
);
7735 cnt
= ring_buffer_entries_cpu(trace_buf
->buffer
, cpu
);
7736 trace_seq_printf(s
, "entries: %ld\n", cnt
);
7738 cnt
= ring_buffer_overrun_cpu(trace_buf
->buffer
, cpu
);
7739 trace_seq_printf(s
, "overrun: %ld\n", cnt
);
7741 cnt
= ring_buffer_commit_overrun_cpu(trace_buf
->buffer
, cpu
);
7742 trace_seq_printf(s
, "commit overrun: %ld\n", cnt
);
7744 cnt
= ring_buffer_bytes_cpu(trace_buf
->buffer
, cpu
);
7745 trace_seq_printf(s
, "bytes: %ld\n", cnt
);
7747 if (trace_clocks
[tr
->clock_id
].in_ns
) {
7748 /* local or global for trace_clock */
7749 t
= ns2usecs(ring_buffer_oldest_event_ts(trace_buf
->buffer
, cpu
));
7750 usec_rem
= do_div(t
, USEC_PER_SEC
);
7751 trace_seq_printf(s
, "oldest event ts: %5llu.%06lu\n",
7754 t
= ns2usecs(ring_buffer_time_stamp(trace_buf
->buffer
, cpu
));
7755 usec_rem
= do_div(t
, USEC_PER_SEC
);
7756 trace_seq_printf(s
, "now ts: %5llu.%06lu\n", t
, usec_rem
);
7758 /* counter or tsc mode for trace_clock */
7759 trace_seq_printf(s
, "oldest event ts: %llu\n",
7760 ring_buffer_oldest_event_ts(trace_buf
->buffer
, cpu
));
7762 trace_seq_printf(s
, "now ts: %llu\n",
7763 ring_buffer_time_stamp(trace_buf
->buffer
, cpu
));
7766 cnt
= ring_buffer_dropped_events_cpu(trace_buf
->buffer
, cpu
);
7767 trace_seq_printf(s
, "dropped events: %ld\n", cnt
);
7769 cnt
= ring_buffer_read_events_cpu(trace_buf
->buffer
, cpu
);
7770 trace_seq_printf(s
, "read events: %ld\n", cnt
);
7772 count
= simple_read_from_buffer(ubuf
, count
, ppos
,
7773 s
->buffer
, trace_seq_used(s
));
7780 static const struct file_operations tracing_stats_fops
= {
7781 .open
= tracing_open_generic_tr
,
7782 .read
= tracing_stats_read
,
7783 .llseek
= generic_file_llseek
,
7784 .release
= tracing_release_generic_tr
,
7787 #ifdef CONFIG_DYNAMIC_FTRACE
7790 tracing_read_dyn_info(struct file
*filp
, char __user
*ubuf
,
7791 size_t cnt
, loff_t
*ppos
)
7797 /* 256 should be plenty to hold the amount needed */
7798 buf
= kmalloc(256, GFP_KERNEL
);
7802 r
= scnprintf(buf
, 256, "%ld pages:%ld groups: %ld\n",
7803 ftrace_update_tot_cnt
,
7804 ftrace_number_of_pages
,
7805 ftrace_number_of_groups
);
7807 ret
= simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, r
);
7812 static const struct file_operations tracing_dyn_info_fops
= {
7813 .open
= tracing_open_generic
,
7814 .read
= tracing_read_dyn_info
,
7815 .llseek
= generic_file_llseek
,
7817 #endif /* CONFIG_DYNAMIC_FTRACE */
7819 #if defined(CONFIG_TRACER_SNAPSHOT) && defined(CONFIG_DYNAMIC_FTRACE)
7821 ftrace_snapshot(unsigned long ip
, unsigned long parent_ip
,
7822 struct trace_array
*tr
, struct ftrace_probe_ops
*ops
,
7825 tracing_snapshot_instance(tr
);
7829 ftrace_count_snapshot(unsigned long ip
, unsigned long parent_ip
,
7830 struct trace_array
*tr
, struct ftrace_probe_ops
*ops
,
7833 struct ftrace_func_mapper
*mapper
= data
;
7837 count
= (long *)ftrace_func_mapper_find_ip(mapper
, ip
);
7847 tracing_snapshot_instance(tr
);
7851 ftrace_snapshot_print(struct seq_file
*m
, unsigned long ip
,
7852 struct ftrace_probe_ops
*ops
, void *data
)
7854 struct ftrace_func_mapper
*mapper
= data
;
7857 seq_printf(m
, "%ps:", (void *)ip
);
7859 seq_puts(m
, "snapshot");
7862 count
= (long *)ftrace_func_mapper_find_ip(mapper
, ip
);
7865 seq_printf(m
, ":count=%ld\n", *count
);
7867 seq_puts(m
, ":unlimited\n");
7873 ftrace_snapshot_init(struct ftrace_probe_ops
*ops
, struct trace_array
*tr
,
7874 unsigned long ip
, void *init_data
, void **data
)
7876 struct ftrace_func_mapper
*mapper
= *data
;
7879 mapper
= allocate_ftrace_func_mapper();
7885 return ftrace_func_mapper_add_ip(mapper
, ip
, init_data
);
7889 ftrace_snapshot_free(struct ftrace_probe_ops
*ops
, struct trace_array
*tr
,
7890 unsigned long ip
, void *data
)
7892 struct ftrace_func_mapper
*mapper
= data
;
7897 free_ftrace_func_mapper(mapper
, NULL
);
7901 ftrace_func_mapper_remove_ip(mapper
, ip
);
7904 static struct ftrace_probe_ops snapshot_probe_ops
= {
7905 .func
= ftrace_snapshot
,
7906 .print
= ftrace_snapshot_print
,
7909 static struct ftrace_probe_ops snapshot_count_probe_ops
= {
7910 .func
= ftrace_count_snapshot
,
7911 .print
= ftrace_snapshot_print
,
7912 .init
= ftrace_snapshot_init
,
7913 .free
= ftrace_snapshot_free
,
7917 ftrace_trace_snapshot_callback(struct trace_array
*tr
, struct ftrace_hash
*hash
,
7918 char *glob
, char *cmd
, char *param
, int enable
)
7920 struct ftrace_probe_ops
*ops
;
7921 void *count
= (void *)-1;
7928 /* hash funcs only work with set_ftrace_filter */
7932 ops
= param
? &snapshot_count_probe_ops
: &snapshot_probe_ops
;
7935 return unregister_ftrace_function_probe_func(glob
+1, tr
, ops
);
7940 number
= strsep(¶m
, ":");
7942 if (!strlen(number
))
7946 * We use the callback data field (which is a pointer)
7949 ret
= kstrtoul(number
, 0, (unsigned long *)&count
);
7954 ret
= tracing_alloc_snapshot_instance(tr
);
7958 ret
= register_ftrace_function_probe(glob
, tr
, ops
, count
);
7961 return ret
< 0 ? ret
: 0;
7964 static struct ftrace_func_command ftrace_snapshot_cmd
= {
7966 .func
= ftrace_trace_snapshot_callback
,
7969 static __init
int register_snapshot_cmd(void)
7971 return register_ftrace_command(&ftrace_snapshot_cmd
);
7974 static inline __init
int register_snapshot_cmd(void) { return 0; }
7975 #endif /* defined(CONFIG_TRACER_SNAPSHOT) && defined(CONFIG_DYNAMIC_FTRACE) */
7977 static struct dentry
*tracing_get_dentry(struct trace_array
*tr
)
7979 if (WARN_ON(!tr
->dir
))
7980 return ERR_PTR(-ENODEV
);
7982 /* Top directory uses NULL as the parent */
7983 if (tr
->flags
& TRACE_ARRAY_FL_GLOBAL
)
7986 /* All sub buffers have a descriptor */
7990 static struct dentry
*tracing_dentry_percpu(struct trace_array
*tr
, int cpu
)
7992 struct dentry
*d_tracer
;
7995 return tr
->percpu_dir
;
7997 d_tracer
= tracing_get_dentry(tr
);
7998 if (IS_ERR(d_tracer
))
8001 tr
->percpu_dir
= tracefs_create_dir("per_cpu", d_tracer
);
8003 MEM_FAIL(!tr
->percpu_dir
,
8004 "Could not create tracefs directory 'per_cpu/%d'\n", cpu
);
8006 return tr
->percpu_dir
;
8009 static struct dentry
*
8010 trace_create_cpu_file(const char *name
, umode_t mode
, struct dentry
*parent
,
8011 void *data
, long cpu
, const struct file_operations
*fops
)
8013 struct dentry
*ret
= trace_create_file(name
, mode
, parent
, data
, fops
);
8015 if (ret
) /* See tracing_get_cpu() */
8016 d_inode(ret
)->i_cdev
= (void *)(cpu
+ 1);
8021 tracing_init_tracefs_percpu(struct trace_array
*tr
, long cpu
)
8023 struct dentry
*d_percpu
= tracing_dentry_percpu(tr
, cpu
);
8024 struct dentry
*d_cpu
;
8025 char cpu_dir
[30]; /* 30 characters should be more than enough */
8030 snprintf(cpu_dir
, 30, "cpu%ld", cpu
);
8031 d_cpu
= tracefs_create_dir(cpu_dir
, d_percpu
);
8033 pr_warn("Could not create tracefs '%s' entry\n", cpu_dir
);
8037 /* per cpu trace_pipe */
8038 trace_create_cpu_file("trace_pipe", 0444, d_cpu
,
8039 tr
, cpu
, &tracing_pipe_fops
);
8042 trace_create_cpu_file("trace", 0644, d_cpu
,
8043 tr
, cpu
, &tracing_fops
);
8045 trace_create_cpu_file("trace_pipe_raw", 0444, d_cpu
,
8046 tr
, cpu
, &tracing_buffers_fops
);
8048 trace_create_cpu_file("stats", 0444, d_cpu
,
8049 tr
, cpu
, &tracing_stats_fops
);
8051 trace_create_cpu_file("buffer_size_kb", 0444, d_cpu
,
8052 tr
, cpu
, &tracing_entries_fops
);
8054 #ifdef CONFIG_TRACER_SNAPSHOT
8055 trace_create_cpu_file("snapshot", 0644, d_cpu
,
8056 tr
, cpu
, &snapshot_fops
);
8058 trace_create_cpu_file("snapshot_raw", 0444, d_cpu
,
8059 tr
, cpu
, &snapshot_raw_fops
);
8063 #ifdef CONFIG_FTRACE_SELFTEST
8064 /* Let selftest have access to static functions in this file */
8065 #include "trace_selftest.c"
8069 trace_options_read(struct file
*filp
, char __user
*ubuf
, size_t cnt
,
8072 struct trace_option_dentry
*topt
= filp
->private_data
;
8075 if (topt
->flags
->val
& topt
->opt
->bit
)
8080 return simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, 2);
8084 trace_options_write(struct file
*filp
, const char __user
*ubuf
, size_t cnt
,
8087 struct trace_option_dentry
*topt
= filp
->private_data
;
8091 ret
= kstrtoul_from_user(ubuf
, cnt
, 10, &val
);
8095 if (val
!= 0 && val
!= 1)
8098 if (!!(topt
->flags
->val
& topt
->opt
->bit
) != val
) {
8099 mutex_lock(&trace_types_lock
);
8100 ret
= __set_tracer_option(topt
->tr
, topt
->flags
,
8102 mutex_unlock(&trace_types_lock
);
8113 static const struct file_operations trace_options_fops
= {
8114 .open
= tracing_open_generic
,
8115 .read
= trace_options_read
,
8116 .write
= trace_options_write
,
8117 .llseek
= generic_file_llseek
,
8121 * In order to pass in both the trace_array descriptor as well as the index
8122 * to the flag that the trace option file represents, the trace_array
8123 * has a character array of trace_flags_index[], which holds the index
8124 * of the bit for the flag it represents. index[0] == 0, index[1] == 1, etc.
8125 * The address of this character array is passed to the flag option file
8126 * read/write callbacks.
8128 * In order to extract both the index and the trace_array descriptor,
8129 * get_tr_index() uses the following algorithm.
8133 * As the pointer itself contains the address of the index (remember
8136 * Then to get the trace_array descriptor, by subtracting that index
8137 * from the ptr, we get to the start of the index itself.
8139 * ptr - idx == &index[0]
8141 * Then a simple container_of() from that pointer gets us to the
8142 * trace_array descriptor.
8144 static void get_tr_index(void *data
, struct trace_array
**ptr
,
8145 unsigned int *pindex
)
8147 *pindex
= *(unsigned char *)data
;
8149 *ptr
= container_of(data
- *pindex
, struct trace_array
,
8154 trace_options_core_read(struct file
*filp
, char __user
*ubuf
, size_t cnt
,
8157 void *tr_index
= filp
->private_data
;
8158 struct trace_array
*tr
;
8162 get_tr_index(tr_index
, &tr
, &index
);
8164 if (tr
->trace_flags
& (1 << index
))
8169 return simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, 2);
8173 trace_options_core_write(struct file
*filp
, const char __user
*ubuf
, size_t cnt
,
8176 void *tr_index
= filp
->private_data
;
8177 struct trace_array
*tr
;
8182 get_tr_index(tr_index
, &tr
, &index
);
8184 ret
= kstrtoul_from_user(ubuf
, cnt
, 10, &val
);
8188 if (val
!= 0 && val
!= 1)
8191 mutex_lock(&event_mutex
);
8192 mutex_lock(&trace_types_lock
);
8193 ret
= set_tracer_flag(tr
, 1 << index
, val
);
8194 mutex_unlock(&trace_types_lock
);
8195 mutex_unlock(&event_mutex
);
8205 static const struct file_operations trace_options_core_fops
= {
8206 .open
= tracing_open_generic
,
8207 .read
= trace_options_core_read
,
8208 .write
= trace_options_core_write
,
8209 .llseek
= generic_file_llseek
,
8212 struct dentry
*trace_create_file(const char *name
,
8214 struct dentry
*parent
,
8216 const struct file_operations
*fops
)
8220 ret
= tracefs_create_file(name
, mode
, parent
, data
, fops
);
8222 pr_warn("Could not create tracefs '%s' entry\n", name
);
8228 static struct dentry
*trace_options_init_dentry(struct trace_array
*tr
)
8230 struct dentry
*d_tracer
;
8235 d_tracer
= tracing_get_dentry(tr
);
8236 if (IS_ERR(d_tracer
))
8239 tr
->options
= tracefs_create_dir("options", d_tracer
);
8241 pr_warn("Could not create tracefs directory 'options'\n");
8249 create_trace_option_file(struct trace_array
*tr
,
8250 struct trace_option_dentry
*topt
,
8251 struct tracer_flags
*flags
,
8252 struct tracer_opt
*opt
)
8254 struct dentry
*t_options
;
8256 t_options
= trace_options_init_dentry(tr
);
8260 topt
->flags
= flags
;
8264 topt
->entry
= trace_create_file(opt
->name
, 0644, t_options
, topt
,
8265 &trace_options_fops
);
8270 create_trace_option_files(struct trace_array
*tr
, struct tracer
*tracer
)
8272 struct trace_option_dentry
*topts
;
8273 struct trace_options
*tr_topts
;
8274 struct tracer_flags
*flags
;
8275 struct tracer_opt
*opts
;
8282 flags
= tracer
->flags
;
8284 if (!flags
|| !flags
->opts
)
8288 * If this is an instance, only create flags for tracers
8289 * the instance may have.
8291 if (!trace_ok_for_array(tracer
, tr
))
8294 for (i
= 0; i
< tr
->nr_topts
; i
++) {
8295 /* Make sure there's no duplicate flags. */
8296 if (WARN_ON_ONCE(tr
->topts
[i
].tracer
->flags
== tracer
->flags
))
8302 for (cnt
= 0; opts
[cnt
].name
; cnt
++)
8305 topts
= kcalloc(cnt
+ 1, sizeof(*topts
), GFP_KERNEL
);
8309 tr_topts
= krealloc(tr
->topts
, sizeof(*tr
->topts
) * (tr
->nr_topts
+ 1),
8316 tr
->topts
= tr_topts
;
8317 tr
->topts
[tr
->nr_topts
].tracer
= tracer
;
8318 tr
->topts
[tr
->nr_topts
].topts
= topts
;
8321 for (cnt
= 0; opts
[cnt
].name
; cnt
++) {
8322 create_trace_option_file(tr
, &topts
[cnt
], flags
,
8324 MEM_FAIL(topts
[cnt
].entry
== NULL
,
8325 "Failed to create trace option: %s",
8330 static struct dentry
*
8331 create_trace_option_core_file(struct trace_array
*tr
,
8332 const char *option
, long index
)
8334 struct dentry
*t_options
;
8336 t_options
= trace_options_init_dentry(tr
);
8340 return trace_create_file(option
, 0644, t_options
,
8341 (void *)&tr
->trace_flags_index
[index
],
8342 &trace_options_core_fops
);
8345 static void create_trace_options_dir(struct trace_array
*tr
)
8347 struct dentry
*t_options
;
8348 bool top_level
= tr
== &global_trace
;
8351 t_options
= trace_options_init_dentry(tr
);
8355 for (i
= 0; trace_options
[i
]; i
++) {
8357 !((1 << i
) & TOP_LEVEL_TRACE_FLAGS
))
8358 create_trace_option_core_file(tr
, trace_options
[i
], i
);
8363 rb_simple_read(struct file
*filp
, char __user
*ubuf
,
8364 size_t cnt
, loff_t
*ppos
)
8366 struct trace_array
*tr
= filp
->private_data
;
8370 r
= tracer_tracing_is_on(tr
);
8371 r
= sprintf(buf
, "%d\n", r
);
8373 return simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, r
);
8377 rb_simple_write(struct file
*filp
, const char __user
*ubuf
,
8378 size_t cnt
, loff_t
*ppos
)
8380 struct trace_array
*tr
= filp
->private_data
;
8381 struct trace_buffer
*buffer
= tr
->array_buffer
.buffer
;
8385 ret
= kstrtoul_from_user(ubuf
, cnt
, 10, &val
);
8390 mutex_lock(&trace_types_lock
);
8391 if (!!val
== tracer_tracing_is_on(tr
)) {
8392 val
= 0; /* do nothing */
8394 tracer_tracing_on(tr
);
8395 if (tr
->current_trace
->start
)
8396 tr
->current_trace
->start(tr
);
8398 tracer_tracing_off(tr
);
8399 if (tr
->current_trace
->stop
)
8400 tr
->current_trace
->stop(tr
);
8402 mutex_unlock(&trace_types_lock
);
8410 static const struct file_operations rb_simple_fops
= {
8411 .open
= tracing_open_generic_tr
,
8412 .read
= rb_simple_read
,
8413 .write
= rb_simple_write
,
8414 .release
= tracing_release_generic_tr
,
8415 .llseek
= default_llseek
,
8419 buffer_percent_read(struct file
*filp
, char __user
*ubuf
,
8420 size_t cnt
, loff_t
*ppos
)
8422 struct trace_array
*tr
= filp
->private_data
;
8426 r
= tr
->buffer_percent
;
8427 r
= sprintf(buf
, "%d\n", r
);
8429 return simple_read_from_buffer(ubuf
, cnt
, ppos
, buf
, r
);
8433 buffer_percent_write(struct file
*filp
, const char __user
*ubuf
,
8434 size_t cnt
, loff_t
*ppos
)
8436 struct trace_array
*tr
= filp
->private_data
;
8440 ret
= kstrtoul_from_user(ubuf
, cnt
, 10, &val
);
8450 tr
->buffer_percent
= val
;
8457 static const struct file_operations buffer_percent_fops
= {
8458 .open
= tracing_open_generic_tr
,
8459 .read
= buffer_percent_read
,
8460 .write
= buffer_percent_write
,
8461 .release
= tracing_release_generic_tr
,
8462 .llseek
= default_llseek
,
8465 static struct dentry
*trace_instance_dir
;
8468 init_tracer_tracefs(struct trace_array
*tr
, struct dentry
*d_tracer
);
8471 allocate_trace_buffer(struct trace_array
*tr
, struct array_buffer
*buf
, int size
)
8473 enum ring_buffer_flags rb_flags
;
8475 rb_flags
= tr
->trace_flags
& TRACE_ITER_OVERWRITE
? RB_FL_OVERWRITE
: 0;
8479 buf
->buffer
= ring_buffer_alloc(size
, rb_flags
);
8483 buf
->data
= alloc_percpu(struct trace_array_cpu
);
8485 ring_buffer_free(buf
->buffer
);
8490 /* Allocate the first page for all buffers */
8491 set_buffer_entries(&tr
->array_buffer
,
8492 ring_buffer_size(tr
->array_buffer
.buffer
, 0));
8497 static int allocate_trace_buffers(struct trace_array
*tr
, int size
)
8501 ret
= allocate_trace_buffer(tr
, &tr
->array_buffer
, size
);
8505 #ifdef CONFIG_TRACER_MAX_TRACE
8506 ret
= allocate_trace_buffer(tr
, &tr
->max_buffer
,
8507 allocate_snapshot
? size
: 1);
8508 if (MEM_FAIL(ret
, "Failed to allocate trace buffer\n")) {
8509 ring_buffer_free(tr
->array_buffer
.buffer
);
8510 tr
->array_buffer
.buffer
= NULL
;
8511 free_percpu(tr
->array_buffer
.data
);
8512 tr
->array_buffer
.data
= NULL
;
8515 tr
->allocated_snapshot
= allocate_snapshot
;
8518 * Only the top level trace array gets its snapshot allocated
8519 * from the kernel command line.
8521 allocate_snapshot
= false;
8527 static void free_trace_buffer(struct array_buffer
*buf
)
8530 ring_buffer_free(buf
->buffer
);
8532 free_percpu(buf
->data
);
8537 static void free_trace_buffers(struct trace_array
*tr
)
8542 free_trace_buffer(&tr
->array_buffer
);
8544 #ifdef CONFIG_TRACER_MAX_TRACE
8545 free_trace_buffer(&tr
->max_buffer
);
8549 static void init_trace_flags_index(struct trace_array
*tr
)
8553 /* Used by the trace options files */
8554 for (i
= 0; i
< TRACE_FLAGS_MAX_SIZE
; i
++)
8555 tr
->trace_flags_index
[i
] = i
;
8558 static void __update_tracer_options(struct trace_array
*tr
)
8562 for (t
= trace_types
; t
; t
= t
->next
)
8563 add_tracer_options(tr
, t
);
8566 static void update_tracer_options(struct trace_array
*tr
)
8568 mutex_lock(&trace_types_lock
);
8569 __update_tracer_options(tr
);
8570 mutex_unlock(&trace_types_lock
);
8573 /* Must have trace_types_lock held */
8574 struct trace_array
*trace_array_find(const char *instance
)
8576 struct trace_array
*tr
, *found
= NULL
;
8578 list_for_each_entry(tr
, &ftrace_trace_arrays
, list
) {
8579 if (tr
->name
&& strcmp(tr
->name
, instance
) == 0) {
8588 struct trace_array
*trace_array_find_get(const char *instance
)
8590 struct trace_array
*tr
;
8592 mutex_lock(&trace_types_lock
);
8593 tr
= trace_array_find(instance
);
8596 mutex_unlock(&trace_types_lock
);
8601 static struct trace_array
*trace_array_create(const char *name
)
8603 struct trace_array
*tr
;
8607 tr
= kzalloc(sizeof(*tr
), GFP_KERNEL
);
8609 return ERR_PTR(ret
);
8611 tr
->name
= kstrdup(name
, GFP_KERNEL
);
8615 if (!alloc_cpumask_var(&tr
->tracing_cpumask
, GFP_KERNEL
))
8618 tr
->trace_flags
= global_trace
.trace_flags
& ~ZEROED_TRACE_FLAGS
;
8620 cpumask_copy(tr
->tracing_cpumask
, cpu_all_mask
);
8622 raw_spin_lock_init(&tr
->start_lock
);
8624 tr
->max_lock
= (arch_spinlock_t
)__ARCH_SPIN_LOCK_UNLOCKED
;
8626 tr
->current_trace
= &nop_trace
;
8628 INIT_LIST_HEAD(&tr
->systems
);
8629 INIT_LIST_HEAD(&tr
->events
);
8630 INIT_LIST_HEAD(&tr
->hist_vars
);
8631 INIT_LIST_HEAD(&tr
->err_log
);
8633 if (allocate_trace_buffers(tr
, trace_buf_size
) < 0)
8636 tr
->dir
= tracefs_create_dir(name
, trace_instance_dir
);
8640 ret
= event_trace_add_tracer(tr
->dir
, tr
);
8642 tracefs_remove(tr
->dir
);
8646 ftrace_init_trace_array(tr
);
8648 init_tracer_tracefs(tr
, tr
->dir
);
8649 init_trace_flags_index(tr
);
8650 __update_tracer_options(tr
);
8652 list_add(&tr
->list
, &ftrace_trace_arrays
);
8660 free_trace_buffers(tr
);
8661 free_cpumask_var(tr
->tracing_cpumask
);
8665 return ERR_PTR(ret
);
8668 static int instance_mkdir(const char *name
)
8670 struct trace_array
*tr
;
8673 mutex_lock(&event_mutex
);
8674 mutex_lock(&trace_types_lock
);
8677 if (trace_array_find(name
))
8680 tr
= trace_array_create(name
);
8682 ret
= PTR_ERR_OR_ZERO(tr
);
8685 mutex_unlock(&trace_types_lock
);
8686 mutex_unlock(&event_mutex
);
8691 * trace_array_get_by_name - Create/Lookup a trace array, given its name.
8692 * @name: The name of the trace array to be looked up/created.
8694 * Returns pointer to trace array with given name.
8695 * NULL, if it cannot be created.
8697 * NOTE: This function increments the reference counter associated with the
8698 * trace array returned. This makes sure it cannot be freed while in use.
8699 * Use trace_array_put() once the trace array is no longer needed.
8700 * If the trace_array is to be freed, trace_array_destroy() needs to
8701 * be called after the trace_array_put(), or simply let user space delete
8702 * it from the tracefs instances directory. But until the
8703 * trace_array_put() is called, user space can not delete it.
8706 struct trace_array
*trace_array_get_by_name(const char *name
)
8708 struct trace_array
*tr
;
8710 mutex_lock(&event_mutex
);
8711 mutex_lock(&trace_types_lock
);
8713 list_for_each_entry(tr
, &ftrace_trace_arrays
, list
) {
8714 if (tr
->name
&& strcmp(tr
->name
, name
) == 0)
8718 tr
= trace_array_create(name
);
8726 mutex_unlock(&trace_types_lock
);
8727 mutex_unlock(&event_mutex
);
8730 EXPORT_SYMBOL_GPL(trace_array_get_by_name
);
8732 static int __remove_instance(struct trace_array
*tr
)
8736 /* Reference counter for a newly created trace array = 1. */
8737 if (tr
->ref
> 1 || (tr
->current_trace
&& tr
->current_trace
->ref
))
8740 list_del(&tr
->list
);
8742 /* Disable all the flags that were enabled coming in */
8743 for (i
= 0; i
< TRACE_FLAGS_MAX_SIZE
; i
++) {
8744 if ((1 << i
) & ZEROED_TRACE_FLAGS
)
8745 set_tracer_flag(tr
, 1 << i
, 0);
8748 tracing_set_nop(tr
);
8749 clear_ftrace_function_probes(tr
);
8750 event_trace_del_tracer(tr
);
8751 ftrace_clear_pids(tr
);
8752 ftrace_destroy_function_files(tr
);
8753 tracefs_remove(tr
->dir
);
8754 free_trace_buffers(tr
);
8756 for (i
= 0; i
< tr
->nr_topts
; i
++) {
8757 kfree(tr
->topts
[i
].topts
);
8761 free_cpumask_var(tr
->tracing_cpumask
);
8769 int trace_array_destroy(struct trace_array
*this_tr
)
8771 struct trace_array
*tr
;
8777 mutex_lock(&event_mutex
);
8778 mutex_lock(&trace_types_lock
);
8782 /* Making sure trace array exists before destroying it. */
8783 list_for_each_entry(tr
, &ftrace_trace_arrays
, list
) {
8784 if (tr
== this_tr
) {
8785 ret
= __remove_instance(tr
);
8790 mutex_unlock(&trace_types_lock
);
8791 mutex_unlock(&event_mutex
);
8795 EXPORT_SYMBOL_GPL(trace_array_destroy
);
8797 static int instance_rmdir(const char *name
)
8799 struct trace_array
*tr
;
8802 mutex_lock(&event_mutex
);
8803 mutex_lock(&trace_types_lock
);
8806 tr
= trace_array_find(name
);
8808 ret
= __remove_instance(tr
);
8810 mutex_unlock(&trace_types_lock
);
8811 mutex_unlock(&event_mutex
);
8816 static __init
void create_trace_instances(struct dentry
*d_tracer
)
8818 trace_instance_dir
= tracefs_create_instance_dir("instances", d_tracer
,
8821 if (MEM_FAIL(!trace_instance_dir
, "Failed to create instances directory\n"))
8826 init_tracer_tracefs(struct trace_array
*tr
, struct dentry
*d_tracer
)
8828 struct trace_event_file
*file
;
8831 trace_create_file("available_tracers", 0444, d_tracer
,
8832 tr
, &show_traces_fops
);
8834 trace_create_file("current_tracer", 0644, d_tracer
,
8835 tr
, &set_tracer_fops
);
8837 trace_create_file("tracing_cpumask", 0644, d_tracer
,
8838 tr
, &tracing_cpumask_fops
);
8840 trace_create_file("trace_options", 0644, d_tracer
,
8841 tr
, &tracing_iter_fops
);
8843 trace_create_file("trace", 0644, d_tracer
,
8846 trace_create_file("trace_pipe", 0444, d_tracer
,
8847 tr
, &tracing_pipe_fops
);
8849 trace_create_file("buffer_size_kb", 0644, d_tracer
,
8850 tr
, &tracing_entries_fops
);
8852 trace_create_file("buffer_total_size_kb", 0444, d_tracer
,
8853 tr
, &tracing_total_entries_fops
);
8855 trace_create_file("free_buffer", 0200, d_tracer
,
8856 tr
, &tracing_free_buffer_fops
);
8858 trace_create_file("trace_marker", 0220, d_tracer
,
8859 tr
, &tracing_mark_fops
);
8861 file
= __find_event_file(tr
, "ftrace", "print");
8862 if (file
&& file
->dir
)
8863 trace_create_file("trigger", 0644, file
->dir
, file
,
8864 &event_trigger_fops
);
8865 tr
->trace_marker_file
= file
;
8867 trace_create_file("trace_marker_raw", 0220, d_tracer
,
8868 tr
, &tracing_mark_raw_fops
);
8870 trace_create_file("trace_clock", 0644, d_tracer
, tr
,
8873 trace_create_file("tracing_on", 0644, d_tracer
,
8874 tr
, &rb_simple_fops
);
8876 trace_create_file("timestamp_mode", 0444, d_tracer
, tr
,
8877 &trace_time_stamp_mode_fops
);
8879 tr
->buffer_percent
= 50;
8881 trace_create_file("buffer_percent", 0444, d_tracer
,
8882 tr
, &buffer_percent_fops
);
8884 create_trace_options_dir(tr
);
8886 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)
8887 trace_create_maxlat_file(tr
, d_tracer
);
8890 if (ftrace_create_function_files(tr
, d_tracer
))
8891 MEM_FAIL(1, "Could not allocate function filter files");
8893 #ifdef CONFIG_TRACER_SNAPSHOT
8894 trace_create_file("snapshot", 0644, d_tracer
,
8895 tr
, &snapshot_fops
);
8898 trace_create_file("error_log", 0644, d_tracer
,
8899 tr
, &tracing_err_log_fops
);
8901 for_each_tracing_cpu(cpu
)
8902 tracing_init_tracefs_percpu(tr
, cpu
);
8904 ftrace_init_tracefs(tr
, d_tracer
);
8907 static struct vfsmount
*trace_automount(struct dentry
*mntpt
, void *ingore
)
8909 struct vfsmount
*mnt
;
8910 struct file_system_type
*type
;
8913 * To maintain backward compatibility for tools that mount
8914 * debugfs to get to the tracing facility, tracefs is automatically
8915 * mounted to the debugfs/tracing directory.
8917 type
= get_fs_type("tracefs");
8920 mnt
= vfs_submount(mntpt
, type
, "tracefs", NULL
);
8921 put_filesystem(type
);
8930 * tracing_init_dentry - initialize top level trace array
8932 * This is called when creating files or directories in the tracing
8933 * directory. It is called via fs_initcall() by any of the boot up code
8934 * and expects to return the dentry of the top level tracing directory.
8936 struct dentry
*tracing_init_dentry(void)
8938 struct trace_array
*tr
= &global_trace
;
8940 if (security_locked_down(LOCKDOWN_TRACEFS
)) {
8941 pr_warn("Tracing disabled due to lockdown\n");
8942 return ERR_PTR(-EPERM
);
8945 /* The top level trace array uses NULL as parent */
8949 if (WARN_ON(!tracefs_initialized()) ||
8950 (IS_ENABLED(CONFIG_DEBUG_FS
) &&
8951 WARN_ON(!debugfs_initialized())))
8952 return ERR_PTR(-ENODEV
);
8955 * As there may still be users that expect the tracing
8956 * files to exist in debugfs/tracing, we must automount
8957 * the tracefs file system there, so older tools still
8958 * work with the newer kerenl.
8960 tr
->dir
= debugfs_create_automount("tracing", NULL
,
8961 trace_automount
, NULL
);
8966 extern struct trace_eval_map
*__start_ftrace_eval_maps
[];
8967 extern struct trace_eval_map
*__stop_ftrace_eval_maps
[];
8969 static void __init
trace_eval_init(void)
8973 len
= __stop_ftrace_eval_maps
- __start_ftrace_eval_maps
;
8974 trace_insert_eval_map(NULL
, __start_ftrace_eval_maps
, len
);
8977 #ifdef CONFIG_MODULES
8978 static void trace_module_add_evals(struct module
*mod
)
8980 if (!mod
->num_trace_evals
)
8984 * Modules with bad taint do not have events created, do
8985 * not bother with enums either.
8987 if (trace_module_has_bad_taint(mod
))
8990 trace_insert_eval_map(mod
, mod
->trace_evals
, mod
->num_trace_evals
);
8993 #ifdef CONFIG_TRACE_EVAL_MAP_FILE
8994 static void trace_module_remove_evals(struct module
*mod
)
8996 union trace_eval_map_item
*map
;
8997 union trace_eval_map_item
**last
= &trace_eval_maps
;
8999 if (!mod
->num_trace_evals
)
9002 mutex_lock(&trace_eval_mutex
);
9004 map
= trace_eval_maps
;
9007 if (map
->head
.mod
== mod
)
9009 map
= trace_eval_jmp_to_tail(map
);
9010 last
= &map
->tail
.next
;
9011 map
= map
->tail
.next
;
9016 *last
= trace_eval_jmp_to_tail(map
)->tail
.next
;
9019 mutex_unlock(&trace_eval_mutex
);
9022 static inline void trace_module_remove_evals(struct module
*mod
) { }
9023 #endif /* CONFIG_TRACE_EVAL_MAP_FILE */
9025 static int trace_module_notify(struct notifier_block
*self
,
9026 unsigned long val
, void *data
)
9028 struct module
*mod
= data
;
9031 case MODULE_STATE_COMING
:
9032 trace_module_add_evals(mod
);
9034 case MODULE_STATE_GOING
:
9035 trace_module_remove_evals(mod
);
9042 static struct notifier_block trace_module_nb
= {
9043 .notifier_call
= trace_module_notify
,
9046 #endif /* CONFIG_MODULES */
9048 static __init
int tracer_init_tracefs(void)
9050 struct dentry
*d_tracer
;
9052 trace_access_lock_init();
9054 d_tracer
= tracing_init_dentry();
9055 if (IS_ERR(d_tracer
))
9060 init_tracer_tracefs(&global_trace
, d_tracer
);
9061 ftrace_init_tracefs_toplevel(&global_trace
, d_tracer
);
9063 trace_create_file("tracing_thresh", 0644, d_tracer
,
9064 &global_trace
, &tracing_thresh_fops
);
9066 trace_create_file("README", 0444, d_tracer
,
9067 NULL
, &tracing_readme_fops
);
9069 trace_create_file("saved_cmdlines", 0444, d_tracer
,
9070 NULL
, &tracing_saved_cmdlines_fops
);
9072 trace_create_file("saved_cmdlines_size", 0644, d_tracer
,
9073 NULL
, &tracing_saved_cmdlines_size_fops
);
9075 trace_create_file("saved_tgids", 0444, d_tracer
,
9076 NULL
, &tracing_saved_tgids_fops
);
9080 trace_create_eval_file(d_tracer
);
9082 #ifdef CONFIG_MODULES
9083 register_module_notifier(&trace_module_nb
);
9086 #ifdef CONFIG_DYNAMIC_FTRACE
9087 trace_create_file("dyn_ftrace_total_info", 0444, d_tracer
,
9088 NULL
, &tracing_dyn_info_fops
);
9091 create_trace_instances(d_tracer
);
9093 update_tracer_options(&global_trace
);
9098 static int trace_panic_handler(struct notifier_block
*this,
9099 unsigned long event
, void *unused
)
9101 if (ftrace_dump_on_oops
)
9102 ftrace_dump(ftrace_dump_on_oops
);
9106 static struct notifier_block trace_panic_notifier
= {
9107 .notifier_call
= trace_panic_handler
,
9109 .priority
= 150 /* priority: INT_MAX >= x >= 0 */
9112 static int trace_die_handler(struct notifier_block
*self
,
9118 if (ftrace_dump_on_oops
)
9119 ftrace_dump(ftrace_dump_on_oops
);
9127 static struct notifier_block trace_die_notifier
= {
9128 .notifier_call
= trace_die_handler
,
9133 * printk is set to max of 1024, we really don't need it that big.
9134 * Nothing should be printing 1000 characters anyway.
9136 #define TRACE_MAX_PRINT 1000
9139 * Define here KERN_TRACE so that we have one place to modify
9140 * it if we decide to change what log level the ftrace dump
9143 #define KERN_TRACE KERN_EMERG
9146 trace_printk_seq(struct trace_seq
*s
)
9148 /* Probably should print a warning here. */
9149 if (s
->seq
.len
>= TRACE_MAX_PRINT
)
9150 s
->seq
.len
= TRACE_MAX_PRINT
;
9153 * More paranoid code. Although the buffer size is set to
9154 * PAGE_SIZE, and TRACE_MAX_PRINT is 1000, this is just
9155 * an extra layer of protection.
9157 if (WARN_ON_ONCE(s
->seq
.len
>= s
->seq
.size
))
9158 s
->seq
.len
= s
->seq
.size
- 1;
9160 /* should be zero ended, but we are paranoid. */
9161 s
->buffer
[s
->seq
.len
] = 0;
9163 printk(KERN_TRACE
"%s", s
->buffer
);
9168 void trace_init_global_iter(struct trace_iterator
*iter
)
9170 iter
->tr
= &global_trace
;
9171 iter
->trace
= iter
->tr
->current_trace
;
9172 iter
->cpu_file
= RING_BUFFER_ALL_CPUS
;
9173 iter
->array_buffer
= &global_trace
.array_buffer
;
9175 if (iter
->trace
&& iter
->trace
->open
)
9176 iter
->trace
->open(iter
);
9178 /* Annotate start of buffers if we had overruns */
9179 if (ring_buffer_overruns(iter
->array_buffer
->buffer
))
9180 iter
->iter_flags
|= TRACE_FILE_ANNOTATE
;
9182 /* Output in nanoseconds only if we are using a clock in nanoseconds. */
9183 if (trace_clocks
[iter
->tr
->clock_id
].in_ns
)
9184 iter
->iter_flags
|= TRACE_FILE_TIME_IN_NS
;
9187 void ftrace_dump(enum ftrace_dump_mode oops_dump_mode
)
9189 /* use static because iter can be a bit big for the stack */
9190 static struct trace_iterator iter
;
9191 static atomic_t dump_running
;
9192 struct trace_array
*tr
= &global_trace
;
9193 unsigned int old_userobj
;
9194 unsigned long flags
;
9197 /* Only allow one dump user at a time. */
9198 if (atomic_inc_return(&dump_running
) != 1) {
9199 atomic_dec(&dump_running
);
9204 * Always turn off tracing when we dump.
9205 * We don't need to show trace output of what happens
9206 * between multiple crashes.
9208 * If the user does a sysrq-z, then they can re-enable
9209 * tracing with echo 1 > tracing_on.
9213 local_irq_save(flags
);
9214 printk_nmi_direct_enter();
9216 /* Simulate the iterator */
9217 trace_init_global_iter(&iter
);
9218 /* Can not use kmalloc for iter.temp */
9219 iter
.temp
= static_temp_buf
;
9220 iter
.temp_size
= STATIC_TEMP_BUF_SIZE
;
9222 for_each_tracing_cpu(cpu
) {
9223 atomic_inc(&per_cpu_ptr(iter
.array_buffer
->data
, cpu
)->disabled
);
9226 old_userobj
= tr
->trace_flags
& TRACE_ITER_SYM_USEROBJ
;
9228 /* don't look at user memory in panic mode */
9229 tr
->trace_flags
&= ~TRACE_ITER_SYM_USEROBJ
;
9231 switch (oops_dump_mode
) {
9233 iter
.cpu_file
= RING_BUFFER_ALL_CPUS
;
9236 iter
.cpu_file
= raw_smp_processor_id();
9241 printk(KERN_TRACE
"Bad dumping mode, switching to all CPUs dump\n");
9242 iter
.cpu_file
= RING_BUFFER_ALL_CPUS
;
9245 printk(KERN_TRACE
"Dumping ftrace buffer:\n");
9247 /* Did function tracer already get disabled? */
9248 if (ftrace_is_dead()) {
9249 printk("# WARNING: FUNCTION TRACING IS CORRUPTED\n");
9250 printk("# MAY BE MISSING FUNCTION EVENTS\n");
9254 * We need to stop all tracing on all CPUS to read the
9255 * the next buffer. This is a bit expensive, but is
9256 * not done often. We fill all what we can read,
9257 * and then release the locks again.
9260 while (!trace_empty(&iter
)) {
9263 printk(KERN_TRACE
"---------------------------------\n");
9267 trace_iterator_reset(&iter
);
9268 iter
.iter_flags
|= TRACE_FILE_LAT_FMT
;
9270 if (trace_find_next_entry_inc(&iter
) != NULL
) {
9273 ret
= print_trace_line(&iter
);
9274 if (ret
!= TRACE_TYPE_NO_CONSUME
)
9275 trace_consume(&iter
);
9277 touch_nmi_watchdog();
9279 trace_printk_seq(&iter
.seq
);
9283 printk(KERN_TRACE
" (ftrace buffer empty)\n");
9285 printk(KERN_TRACE
"---------------------------------\n");
9288 tr
->trace_flags
|= old_userobj
;
9290 for_each_tracing_cpu(cpu
) {
9291 atomic_dec(&per_cpu_ptr(iter
.array_buffer
->data
, cpu
)->disabled
);
9293 atomic_dec(&dump_running
);
9294 printk_nmi_direct_exit();
9295 local_irq_restore(flags
);
9297 EXPORT_SYMBOL_GPL(ftrace_dump
);
9299 int trace_run_command(const char *buf
, int (*createfn
)(int, char **))
9306 argv
= argv_split(GFP_KERNEL
, buf
, &argc
);
9311 ret
= createfn(argc
, argv
);
9318 #define WRITE_BUFSIZE 4096
9320 ssize_t
trace_parse_run_command(struct file
*file
, const char __user
*buffer
,
9321 size_t count
, loff_t
*ppos
,
9322 int (*createfn
)(int, char **))
9324 char *kbuf
, *buf
, *tmp
;
9329 kbuf
= kmalloc(WRITE_BUFSIZE
, GFP_KERNEL
);
9333 while (done
< count
) {
9334 size
= count
- done
;
9336 if (size
>= WRITE_BUFSIZE
)
9337 size
= WRITE_BUFSIZE
- 1;
9339 if (copy_from_user(kbuf
, buffer
+ done
, size
)) {
9346 tmp
= strchr(buf
, '\n');
9349 size
= tmp
- buf
+ 1;
9352 if (done
+ size
< count
) {
9355 /* This can accept WRITE_BUFSIZE - 2 ('\n' + '\0') */
9356 pr_warn("Line length is too long: Should be less than %d\n",
9364 /* Remove comments */
9365 tmp
= strchr(buf
, '#');
9370 ret
= trace_run_command(buf
, createfn
);
9375 } while (done
< count
);
9385 __init
static int tracer_alloc_buffers(void)
9391 if (security_locked_down(LOCKDOWN_TRACEFS
)) {
9392 pr_warn("Tracing disabled due to lockdown\n");
9397 * Make sure we don't accidently add more trace options
9398 * than we have bits for.
9400 BUILD_BUG_ON(TRACE_ITER_LAST_BIT
> TRACE_FLAGS_MAX_SIZE
);
9402 if (!alloc_cpumask_var(&tracing_buffer_mask
, GFP_KERNEL
))
9405 if (!alloc_cpumask_var(&global_trace
.tracing_cpumask
, GFP_KERNEL
))
9406 goto out_free_buffer_mask
;
9408 /* Only allocate trace_printk buffers if a trace_printk exists */
9409 if (&__stop___trace_bprintk_fmt
!= &__start___trace_bprintk_fmt
)
9410 /* Must be called before global_trace.buffer is allocated */
9411 trace_printk_init_buffers();
9413 /* To save memory, keep the ring buffer size to its minimum */
9414 if (ring_buffer_expanded
)
9415 ring_buf_size
= trace_buf_size
;
9419 cpumask_copy(tracing_buffer_mask
, cpu_possible_mask
);
9420 cpumask_copy(global_trace
.tracing_cpumask
, cpu_all_mask
);
9422 raw_spin_lock_init(&global_trace
.start_lock
);
9425 * The prepare callbacks allocates some memory for the ring buffer. We
9426 * don't free the buffer if the if the CPU goes down. If we were to free
9427 * the buffer, then the user would lose any trace that was in the
9428 * buffer. The memory will be removed once the "instance" is removed.
9430 ret
= cpuhp_setup_state_multi(CPUHP_TRACE_RB_PREPARE
,
9431 "trace/RB:preapre", trace_rb_cpu_prepare
,
9434 goto out_free_cpumask
;
9435 /* Used for event triggers */
9437 temp_buffer
= ring_buffer_alloc(PAGE_SIZE
, RB_FL_OVERWRITE
);
9439 goto out_rm_hp_state
;
9441 if (trace_create_savedcmd() < 0)
9442 goto out_free_temp_buffer
;
9444 /* TODO: make the number of buffers hot pluggable with CPUS */
9445 if (allocate_trace_buffers(&global_trace
, ring_buf_size
) < 0) {
9446 MEM_FAIL(1, "tracer: failed to allocate ring buffer!\n");
9447 goto out_free_savedcmd
;
9450 if (global_trace
.buffer_disabled
)
9453 if (trace_boot_clock
) {
9454 ret
= tracing_set_clock(&global_trace
, trace_boot_clock
);
9456 pr_warn("Trace clock %s not defined, going back to default\n",
9461 * register_tracer() might reference current_trace, so it
9462 * needs to be set before we register anything. This is
9463 * just a bootstrap of current_trace anyway.
9465 global_trace
.current_trace
= &nop_trace
;
9467 global_trace
.max_lock
= (arch_spinlock_t
)__ARCH_SPIN_LOCK_UNLOCKED
;
9469 ftrace_init_global_array_ops(&global_trace
);
9471 init_trace_flags_index(&global_trace
);
9473 register_tracer(&nop_trace
);
9475 /* Function tracing may start here (via kernel command line) */
9476 init_function_trace();
9478 /* All seems OK, enable tracing */
9479 tracing_disabled
= 0;
9481 atomic_notifier_chain_register(&panic_notifier_list
,
9482 &trace_panic_notifier
);
9484 register_die_notifier(&trace_die_notifier
);
9486 global_trace
.flags
= TRACE_ARRAY_FL_GLOBAL
;
9488 INIT_LIST_HEAD(&global_trace
.systems
);
9489 INIT_LIST_HEAD(&global_trace
.events
);
9490 INIT_LIST_HEAD(&global_trace
.hist_vars
);
9491 INIT_LIST_HEAD(&global_trace
.err_log
);
9492 list_add(&global_trace
.list
, &ftrace_trace_arrays
);
9494 apply_trace_boot_options();
9496 register_snapshot_cmd();
9501 free_saved_cmdlines_buffer(savedcmd
);
9502 out_free_temp_buffer
:
9503 ring_buffer_free(temp_buffer
);
9505 cpuhp_remove_multi_state(CPUHP_TRACE_RB_PREPARE
);
9507 free_cpumask_var(global_trace
.tracing_cpumask
);
9508 out_free_buffer_mask
:
9509 free_cpumask_var(tracing_buffer_mask
);
9514 void __init
early_trace_init(void)
9516 if (tracepoint_printk
) {
9517 tracepoint_print_iter
=
9518 kmalloc(sizeof(*tracepoint_print_iter
), GFP_KERNEL
);
9519 if (MEM_FAIL(!tracepoint_print_iter
,
9520 "Failed to allocate trace iterator\n"))
9521 tracepoint_printk
= 0;
9523 static_key_enable(&tracepoint_printk_key
.key
);
9525 tracer_alloc_buffers();
9528 void __init
trace_init(void)
9533 __init
static int clear_boot_tracer(void)
9536 * The default tracer at boot buffer is an init section.
9537 * This function is called in lateinit. If we did not
9538 * find the boot tracer, then clear it out, to prevent
9539 * later registration from accessing the buffer that is
9540 * about to be freed.
9542 if (!default_bootup_tracer
)
9545 printk(KERN_INFO
"ftrace bootup tracer '%s' not registered.\n",
9546 default_bootup_tracer
);
9547 default_bootup_tracer
= NULL
;
9552 fs_initcall(tracer_init_tracefs
);
9553 late_initcall_sync(clear_boot_tracer
);
9555 #ifdef CONFIG_HAVE_UNSTABLE_SCHED_CLOCK
9556 __init
static int tracing_set_default_clock(void)
9558 /* sched_clock_stable() is determined in late_initcall */
9559 if (!trace_boot_clock
&& !sched_clock_stable()) {
9560 if (security_locked_down(LOCKDOWN_TRACEFS
)) {
9561 pr_warn("Can not set tracing clock due to lockdown\n");
9566 "Unstable clock detected, switching default tracing clock to \"global\"\n"
9567 "If you want to keep using the local clock, then add:\n"
9568 " \"trace_clock=local\"\n"
9569 "on the kernel command line\n");
9570 tracing_set_clock(&global_trace
, "global");
9575 late_initcall_sync(tracing_set_default_clock
);