1e5a81b62SSteven Rostedt /* 2e5a81b62SSteven Rostedt * Copyright (C) 2008 Steven Rostedt <[email protected]> 3e5a81b62SSteven Rostedt * 4e5a81b62SSteven Rostedt */ 5e5a81b62SSteven Rostedt #include <linux/stacktrace.h> 6e5a81b62SSteven Rostedt #include <linux/kallsyms.h> 7e5a81b62SSteven Rostedt #include <linux/seq_file.h> 8e5a81b62SSteven Rostedt #include <linux/spinlock.h> 9e5a81b62SSteven Rostedt #include <linux/uaccess.h> 10e5a81b62SSteven Rostedt #include <linux/debugfs.h> 11e5a81b62SSteven Rostedt #include <linux/ftrace.h> 12e5a81b62SSteven Rostedt #include <linux/module.h> 13f38f1d2aSSteven Rostedt #include <linux/sysctl.h> 14e5a81b62SSteven Rostedt #include <linux/init.h> 15e5a81b62SSteven Rostedt #include <linux/fs.h> 16762e1207SSteven Rostedt 17762e1207SSteven Rostedt #include <asm/setup.h> 18762e1207SSteven Rostedt 19e5a81b62SSteven Rostedt #include "trace.h" 20e5a81b62SSteven Rostedt 21e5a81b62SSteven Rostedt #define STACK_TRACE_ENTRIES 500 22e5a81b62SSteven Rostedt 23d4ecbfc4SSteven Rostedt (Red Hat) #ifdef CC_USING_FENTRY 244df29712SSteven Rostedt (Red Hat) # define fentry 1 25d4ecbfc4SSteven Rostedt (Red Hat) #else 264df29712SSteven Rostedt (Red Hat) # define fentry 0 27d4ecbfc4SSteven Rostedt (Red Hat) #endif 28d4ecbfc4SSteven Rostedt (Red Hat) 291b6cced6SSteven Rostedt static unsigned long stack_dump_trace[STACK_TRACE_ENTRIES+1] = 301b6cced6SSteven Rostedt { [0 ... (STACK_TRACE_ENTRIES)] = ULONG_MAX }; 311b6cced6SSteven Rostedt static unsigned stack_dump_index[STACK_TRACE_ENTRIES]; 321b6cced6SSteven Rostedt 334df29712SSteven Rostedt (Red Hat) /* 344df29712SSteven Rostedt (Red Hat) * Reserve one entry for the passed in ip. This will allow 354df29712SSteven Rostedt (Red Hat) * us to remove most or all of the stack size overhead 364df29712SSteven Rostedt (Red Hat) * added by the stack tracer itself. 374df29712SSteven Rostedt (Red Hat) */ 38e5a81b62SSteven Rostedt static struct stack_trace max_stack_trace = { 394df29712SSteven Rostedt (Red Hat) .max_entries = STACK_TRACE_ENTRIES - 1, 404df29712SSteven Rostedt (Red Hat) .entries = &stack_dump_trace[1], 41e5a81b62SSteven Rostedt }; 42e5a81b62SSteven Rostedt 43e5a81b62SSteven Rostedt static unsigned long max_stack_size; 44445c8951SThomas Gleixner static arch_spinlock_t max_stack_lock = 45edc35bd7SThomas Gleixner (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED; 46e5a81b62SSteven Rostedt 47e5a81b62SSteven Rostedt static DEFINE_PER_CPU(int, trace_active); 48f38f1d2aSSteven Rostedt static DEFINE_MUTEX(stack_sysctl_mutex); 49f38f1d2aSSteven Rostedt 50f38f1d2aSSteven Rostedt int stack_tracer_enabled; 51f38f1d2aSSteven Rostedt static int last_stack_tracer_enabled; 52e5a81b62SSteven Rostedt 5387889501SSteven Rostedt (Red Hat) static inline void 54d4ecbfc4SSteven Rostedt (Red Hat) check_stack(unsigned long ip, unsigned long *stack) 55e5a81b62SSteven Rostedt { 561b6cced6SSteven Rostedt unsigned long this_size, flags; 571b6cced6SSteven Rostedt unsigned long *p, *top, *start; 584df29712SSteven Rostedt (Red Hat) static int tracer_frame; 594df29712SSteven Rostedt (Red Hat) int frame_size = ACCESS_ONCE(tracer_frame); 601b6cced6SSteven Rostedt int i; 61e5a81b62SSteven Rostedt 6287889501SSteven Rostedt (Red Hat) this_size = ((unsigned long)stack) & (THREAD_SIZE-1); 63e5a81b62SSteven Rostedt this_size = THREAD_SIZE - this_size; 644df29712SSteven Rostedt (Red Hat) /* Remove the frame of the tracer */ 654df29712SSteven Rostedt (Red Hat) this_size -= frame_size; 66e5a81b62SSteven Rostedt 67e5a81b62SSteven Rostedt if (this_size <= max_stack_size) 68e5a81b62SSteven Rostedt return; 69e5a81b62SSteven Rostedt 7081520a1bSSteven Rostedt /* we do not handle interrupt stacks yet */ 7187889501SSteven Rostedt (Red Hat) if (!object_is_on_stack(stack)) 7281520a1bSSteven Rostedt return; 7381520a1bSSteven Rostedt 74a5e25883SSteven Rostedt local_irq_save(flags); 750199c4e6SThomas Gleixner arch_spin_lock(&max_stack_lock); 76e5a81b62SSteven Rostedt 774df29712SSteven Rostedt (Red Hat) /* In case another CPU set the tracer_frame on us */ 784df29712SSteven Rostedt (Red Hat) if (unlikely(!frame_size)) 794df29712SSteven Rostedt (Red Hat) this_size -= tracer_frame; 804df29712SSteven Rostedt (Red Hat) 81e5a81b62SSteven Rostedt /* a race could have already updated it */ 82e5a81b62SSteven Rostedt if (this_size <= max_stack_size) 83e5a81b62SSteven Rostedt goto out; 84e5a81b62SSteven Rostedt 85e5a81b62SSteven Rostedt max_stack_size = this_size; 86e5a81b62SSteven Rostedt 87e5a81b62SSteven Rostedt max_stack_trace.nr_entries = 0; 881b6cced6SSteven Rostedt max_stack_trace.skip = 3; 89e5a81b62SSteven Rostedt 90e5a81b62SSteven Rostedt save_stack_trace(&max_stack_trace); 91e5a81b62SSteven Rostedt 921b6cced6SSteven Rostedt /* 934df29712SSteven Rostedt (Red Hat) * Add the passed in ip from the function tracer. 944df29712SSteven Rostedt (Red Hat) * Searching for this on the stack will skip over 954df29712SSteven Rostedt (Red Hat) * most of the overhead from the stack tracer itself. 96d4ecbfc4SSteven Rostedt (Red Hat) */ 97d4ecbfc4SSteven Rostedt (Red Hat) stack_dump_trace[0] = ip; 98d4ecbfc4SSteven Rostedt (Red Hat) max_stack_trace.nr_entries++; 99d4ecbfc4SSteven Rostedt (Red Hat) 100d4ecbfc4SSteven Rostedt (Red Hat) /* 1011b6cced6SSteven Rostedt * Now find where in the stack these are. 1021b6cced6SSteven Rostedt */ 1031b6cced6SSteven Rostedt i = 0; 10487889501SSteven Rostedt (Red Hat) start = stack; 1051b6cced6SSteven Rostedt top = (unsigned long *) 1061b6cced6SSteven Rostedt (((unsigned long)start & ~(THREAD_SIZE-1)) + THREAD_SIZE); 1071b6cced6SSteven Rostedt 1081b6cced6SSteven Rostedt /* 1091b6cced6SSteven Rostedt * Loop through all the entries. One of the entries may 1101b6cced6SSteven Rostedt * for some reason be missed on the stack, so we may 1111b6cced6SSteven Rostedt * have to account for them. If they are all there, this 1121b6cced6SSteven Rostedt * loop will only happen once. This code only takes place 1131b6cced6SSteven Rostedt * on a new max, so it is far from a fast path. 1141b6cced6SSteven Rostedt */ 1151b6cced6SSteven Rostedt while (i < max_stack_trace.nr_entries) { 1160a37119dSSteven Rostedt int found = 0; 1171b6cced6SSteven Rostedt 1181b6cced6SSteven Rostedt stack_dump_index[i] = this_size; 1191b6cced6SSteven Rostedt p = start; 1201b6cced6SSteven Rostedt 1211b6cced6SSteven Rostedt for (; p < top && i < max_stack_trace.nr_entries; p++) { 1221b6cced6SSteven Rostedt if (*p == stack_dump_trace[i]) { 1231b6cced6SSteven Rostedt this_size = stack_dump_index[i++] = 1241b6cced6SSteven Rostedt (top - p) * sizeof(unsigned long); 1250a37119dSSteven Rostedt found = 1; 1261b6cced6SSteven Rostedt /* Start the search from here */ 1271b6cced6SSteven Rostedt start = p + 1; 1284df29712SSteven Rostedt (Red Hat) /* 1294df29712SSteven Rostedt (Red Hat) * We do not want to show the overhead 1304df29712SSteven Rostedt (Red Hat) * of the stack tracer stack in the 1314df29712SSteven Rostedt (Red Hat) * max stack. If we haven't figured 1324df29712SSteven Rostedt (Red Hat) * out what that is, then figure it out 1334df29712SSteven Rostedt (Red Hat) * now. 1344df29712SSteven Rostedt (Red Hat) */ 1354df29712SSteven Rostedt (Red Hat) if (unlikely(!tracer_frame) && i == 1) { 1364df29712SSteven Rostedt (Red Hat) tracer_frame = (p - stack) * 1374df29712SSteven Rostedt (Red Hat) sizeof(unsigned long); 1384df29712SSteven Rostedt (Red Hat) max_stack_size -= tracer_frame; 1394df29712SSteven Rostedt (Red Hat) } 1401b6cced6SSteven Rostedt } 1411b6cced6SSteven Rostedt } 1421b6cced6SSteven Rostedt 1430a37119dSSteven Rostedt if (!found) 1441b6cced6SSteven Rostedt i++; 1451b6cced6SSteven Rostedt } 1461b6cced6SSteven Rostedt 147e5a81b62SSteven Rostedt out: 1480199c4e6SThomas Gleixner arch_spin_unlock(&max_stack_lock); 149a5e25883SSteven Rostedt local_irq_restore(flags); 150e5a81b62SSteven Rostedt } 151e5a81b62SSteven Rostedt 152e5a81b62SSteven Rostedt static void 153a1e2e31dSSteven Rostedt stack_trace_call(unsigned long ip, unsigned long parent_ip, 154a1e2e31dSSteven Rostedt struct ftrace_ops *op, struct pt_regs *pt_regs) 155e5a81b62SSteven Rostedt { 15687889501SSteven Rostedt (Red Hat) unsigned long stack; 1575168ae50SSteven Rostedt int cpu; 158e5a81b62SSteven Rostedt 1595168ae50SSteven Rostedt preempt_disable_notrace(); 160e5a81b62SSteven Rostedt 161e5a81b62SSteven Rostedt cpu = raw_smp_processor_id(); 162e5a81b62SSteven Rostedt /* no atomic needed, we only modify this variable by this cpu */ 163e5a81b62SSteven Rostedt if (per_cpu(trace_active, cpu)++ != 0) 164e5a81b62SSteven Rostedt goto out; 165e5a81b62SSteven Rostedt 1664df29712SSteven Rostedt (Red Hat) /* 1674df29712SSteven Rostedt (Red Hat) * When fentry is used, the traced function does not get 1684df29712SSteven Rostedt (Red Hat) * its stack frame set up, and we lose the parent. 1694df29712SSteven Rostedt (Red Hat) * The ip is pretty useless because the function tracer 1704df29712SSteven Rostedt (Red Hat) * was called before that function set up its stack frame. 1714df29712SSteven Rostedt (Red Hat) * In this case, we use the parent ip. 1724df29712SSteven Rostedt (Red Hat) * 1734df29712SSteven Rostedt (Red Hat) * By adding the return address of either the parent ip 1744df29712SSteven Rostedt (Red Hat) * or the current ip we can disregard most of the stack usage 1754df29712SSteven Rostedt (Red Hat) * caused by the stack tracer itself. 1764df29712SSteven Rostedt (Red Hat) * 1774df29712SSteven Rostedt (Red Hat) * The function tracer always reports the address of where the 1784df29712SSteven Rostedt (Red Hat) * mcount call was, but the stack will hold the return address. 1794df29712SSteven Rostedt (Red Hat) */ 1804df29712SSteven Rostedt (Red Hat) if (fentry) 1814df29712SSteven Rostedt (Red Hat) ip = parent_ip; 1824df29712SSteven Rostedt (Red Hat) else 1834df29712SSteven Rostedt (Red Hat) ip += MCOUNT_INSN_SIZE; 1844df29712SSteven Rostedt (Red Hat) 1854df29712SSteven Rostedt (Red Hat) check_stack(ip, &stack); 186e5a81b62SSteven Rostedt 187e5a81b62SSteven Rostedt out: 188e5a81b62SSteven Rostedt per_cpu(trace_active, cpu)--; 189e5a81b62SSteven Rostedt /* prevent recursion in schedule */ 1905168ae50SSteven Rostedt preempt_enable_notrace(); 191e5a81b62SSteven Rostedt } 192e5a81b62SSteven Rostedt 193e5a81b62SSteven Rostedt static struct ftrace_ops trace_ops __read_mostly = 194e5a81b62SSteven Rostedt { 195e5a81b62SSteven Rostedt .func = stack_trace_call, 1964740974aSSteven Rostedt .flags = FTRACE_OPS_FL_RECURSION_SAFE, 197e5a81b62SSteven Rostedt }; 198e5a81b62SSteven Rostedt 199e5a81b62SSteven Rostedt static ssize_t 200e5a81b62SSteven Rostedt stack_max_size_read(struct file *filp, char __user *ubuf, 201e5a81b62SSteven Rostedt size_t count, loff_t *ppos) 202e5a81b62SSteven Rostedt { 203e5a81b62SSteven Rostedt unsigned long *ptr = filp->private_data; 204e5a81b62SSteven Rostedt char buf[64]; 205e5a81b62SSteven Rostedt int r; 206e5a81b62SSteven Rostedt 207e5a81b62SSteven Rostedt r = snprintf(buf, sizeof(buf), "%ld\n", *ptr); 208e5a81b62SSteven Rostedt if (r > sizeof(buf)) 209e5a81b62SSteven Rostedt r = sizeof(buf); 210e5a81b62SSteven Rostedt return simple_read_from_buffer(ubuf, count, ppos, buf, r); 211e5a81b62SSteven Rostedt } 212e5a81b62SSteven Rostedt 213e5a81b62SSteven Rostedt static ssize_t 214e5a81b62SSteven Rostedt stack_max_size_write(struct file *filp, const char __user *ubuf, 215e5a81b62SSteven Rostedt size_t count, loff_t *ppos) 216e5a81b62SSteven Rostedt { 217e5a81b62SSteven Rostedt long *ptr = filp->private_data; 218e5a81b62SSteven Rostedt unsigned long val, flags; 219e5a81b62SSteven Rostedt int ret; 2204f48f8b7SLai Jiangshan int cpu; 221e5a81b62SSteven Rostedt 22222fe9b54SPeter Huewe ret = kstrtoul_from_user(ubuf, count, 10, &val); 22322fe9b54SPeter Huewe if (ret) 224e5a81b62SSteven Rostedt return ret; 225e5a81b62SSteven Rostedt 226a5e25883SSteven Rostedt local_irq_save(flags); 2274f48f8b7SLai Jiangshan 2284f48f8b7SLai Jiangshan /* 2294f48f8b7SLai Jiangshan * In case we trace inside arch_spin_lock() or after (NMI), 2304f48f8b7SLai Jiangshan * we will cause circular lock, so we also need to increase 2314f48f8b7SLai Jiangshan * the percpu trace_active here. 2324f48f8b7SLai Jiangshan */ 2334f48f8b7SLai Jiangshan cpu = smp_processor_id(); 2344f48f8b7SLai Jiangshan per_cpu(trace_active, cpu)++; 2354f48f8b7SLai Jiangshan 2360199c4e6SThomas Gleixner arch_spin_lock(&max_stack_lock); 237e5a81b62SSteven Rostedt *ptr = val; 2380199c4e6SThomas Gleixner arch_spin_unlock(&max_stack_lock); 2394f48f8b7SLai Jiangshan 2404f48f8b7SLai Jiangshan per_cpu(trace_active, cpu)--; 241a5e25883SSteven Rostedt local_irq_restore(flags); 242e5a81b62SSteven Rostedt 243e5a81b62SSteven Rostedt return count; 244e5a81b62SSteven Rostedt } 245e5a81b62SSteven Rostedt 246f38f1d2aSSteven Rostedt static const struct file_operations stack_max_size_fops = { 247e5a81b62SSteven Rostedt .open = tracing_open_generic, 248e5a81b62SSteven Rostedt .read = stack_max_size_read, 249e5a81b62SSteven Rostedt .write = stack_max_size_write, 2506038f373SArnd Bergmann .llseek = default_llseek, 251e5a81b62SSteven Rostedt }; 252e5a81b62SSteven Rostedt 253e5a81b62SSteven Rostedt static void * 2542fc5f0cfSLi Zefan __next(struct seq_file *m, loff_t *pos) 255e5a81b62SSteven Rostedt { 2562fc5f0cfSLi Zefan long n = *pos - 1; 257e5a81b62SSteven Rostedt 2582fc5f0cfSLi Zefan if (n >= max_stack_trace.nr_entries || stack_dump_trace[n] == ULONG_MAX) 259e5a81b62SSteven Rostedt return NULL; 260e5a81b62SSteven Rostedt 2612fc5f0cfSLi Zefan m->private = (void *)n; 2621b6cced6SSteven Rostedt return &m->private; 263e5a81b62SSteven Rostedt } 264e5a81b62SSteven Rostedt 2652fc5f0cfSLi Zefan static void * 2662fc5f0cfSLi Zefan t_next(struct seq_file *m, void *v, loff_t *pos) 2672fc5f0cfSLi Zefan { 2682fc5f0cfSLi Zefan (*pos)++; 2692fc5f0cfSLi Zefan return __next(m, pos); 2702fc5f0cfSLi Zefan } 2712fc5f0cfSLi Zefan 272e5a81b62SSteven Rostedt static void *t_start(struct seq_file *m, loff_t *pos) 273e5a81b62SSteven Rostedt { 2744f48f8b7SLai Jiangshan int cpu; 2754f48f8b7SLai Jiangshan 276e5a81b62SSteven Rostedt local_irq_disable(); 2774f48f8b7SLai Jiangshan 2784f48f8b7SLai Jiangshan cpu = smp_processor_id(); 2794f48f8b7SLai Jiangshan per_cpu(trace_active, cpu)++; 2804f48f8b7SLai Jiangshan 2810199c4e6SThomas Gleixner arch_spin_lock(&max_stack_lock); 282e5a81b62SSteven Rostedt 283522a110bSLiming Wang if (*pos == 0) 284522a110bSLiming Wang return SEQ_START_TOKEN; 285522a110bSLiming Wang 2862fc5f0cfSLi Zefan return __next(m, pos); 287e5a81b62SSteven Rostedt } 288e5a81b62SSteven Rostedt 289e5a81b62SSteven Rostedt static void t_stop(struct seq_file *m, void *p) 290e5a81b62SSteven Rostedt { 2914f48f8b7SLai Jiangshan int cpu; 2924f48f8b7SLai Jiangshan 2930199c4e6SThomas Gleixner arch_spin_unlock(&max_stack_lock); 2944f48f8b7SLai Jiangshan 2954f48f8b7SLai Jiangshan cpu = smp_processor_id(); 2964f48f8b7SLai Jiangshan per_cpu(trace_active, cpu)--; 2974f48f8b7SLai Jiangshan 298e5a81b62SSteven Rostedt local_irq_enable(); 299e5a81b62SSteven Rostedt } 300e5a81b62SSteven Rostedt 3011b6cced6SSteven Rostedt static int trace_lookup_stack(struct seq_file *m, long i) 302e5a81b62SSteven Rostedt { 3031b6cced6SSteven Rostedt unsigned long addr = stack_dump_trace[i]; 304e5a81b62SSteven Rostedt 305151772dbSAnton Blanchard return seq_printf(m, "%pS\n", (void *)addr); 306e5a81b62SSteven Rostedt } 307e5a81b62SSteven Rostedt 308e447e1dfSSteven Rostedt static void print_disabled(struct seq_file *m) 309e447e1dfSSteven Rostedt { 310e447e1dfSSteven Rostedt seq_puts(m, "#\n" 311e447e1dfSSteven Rostedt "# Stack tracer disabled\n" 312e447e1dfSSteven Rostedt "#\n" 313e447e1dfSSteven Rostedt "# To enable the stack tracer, either add 'stacktrace' to the\n" 314e447e1dfSSteven Rostedt "# kernel command line\n" 315e447e1dfSSteven Rostedt "# or 'echo 1 > /proc/sys/kernel/stack_tracer_enabled'\n" 316e447e1dfSSteven Rostedt "#\n"); 317e447e1dfSSteven Rostedt } 318e447e1dfSSteven Rostedt 319e5a81b62SSteven Rostedt static int t_show(struct seq_file *m, void *v) 320e5a81b62SSteven Rostedt { 321522a110bSLiming Wang long i; 3221b6cced6SSteven Rostedt int size; 323e5a81b62SSteven Rostedt 324522a110bSLiming Wang if (v == SEQ_START_TOKEN) { 3251b6cced6SSteven Rostedt seq_printf(m, " Depth Size Location" 3261b6cced6SSteven Rostedt " (%d entries)\n" 3271b6cced6SSteven Rostedt " ----- ---- --------\n", 328083a63b4Swalimis max_stack_trace.nr_entries - 1); 329e447e1dfSSteven Rostedt 330e447e1dfSSteven Rostedt if (!stack_tracer_enabled && !max_stack_size) 331e447e1dfSSteven Rostedt print_disabled(m); 332e447e1dfSSteven Rostedt 3331b6cced6SSteven Rostedt return 0; 3341b6cced6SSteven Rostedt } 3351b6cced6SSteven Rostedt 336522a110bSLiming Wang i = *(long *)v; 337522a110bSLiming Wang 3381b6cced6SSteven Rostedt if (i >= max_stack_trace.nr_entries || 3391b6cced6SSteven Rostedt stack_dump_trace[i] == ULONG_MAX) 340e5a81b62SSteven Rostedt return 0; 341e5a81b62SSteven Rostedt 3421b6cced6SSteven Rostedt if (i+1 == max_stack_trace.nr_entries || 3431b6cced6SSteven Rostedt stack_dump_trace[i+1] == ULONG_MAX) 3441b6cced6SSteven Rostedt size = stack_dump_index[i]; 3451b6cced6SSteven Rostedt else 3461b6cced6SSteven Rostedt size = stack_dump_index[i] - stack_dump_index[i+1]; 3471b6cced6SSteven Rostedt 3481b6cced6SSteven Rostedt seq_printf(m, "%3ld) %8d %5d ", i, stack_dump_index[i], size); 3491b6cced6SSteven Rostedt 3501b6cced6SSteven Rostedt trace_lookup_stack(m, i); 351e5a81b62SSteven Rostedt 352e5a81b62SSteven Rostedt return 0; 353e5a81b62SSteven Rostedt } 354e5a81b62SSteven Rostedt 355f38f1d2aSSteven Rostedt static const struct seq_operations stack_trace_seq_ops = { 356e5a81b62SSteven Rostedt .start = t_start, 357e5a81b62SSteven Rostedt .next = t_next, 358e5a81b62SSteven Rostedt .stop = t_stop, 359e5a81b62SSteven Rostedt .show = t_show, 360e5a81b62SSteven Rostedt }; 361e5a81b62SSteven Rostedt 362e5a81b62SSteven Rostedt static int stack_trace_open(struct inode *inode, struct file *file) 363e5a81b62SSteven Rostedt { 364d8cc1ab7SLi Zefan return seq_open(file, &stack_trace_seq_ops); 365e5a81b62SSteven Rostedt } 366e5a81b62SSteven Rostedt 367f38f1d2aSSteven Rostedt static const struct file_operations stack_trace_fops = { 368e5a81b62SSteven Rostedt .open = stack_trace_open, 369e5a81b62SSteven Rostedt .read = seq_read, 370e5a81b62SSteven Rostedt .llseek = seq_lseek, 371d8cc1ab7SLi Zefan .release = seq_release, 372e5a81b62SSteven Rostedt }; 373e5a81b62SSteven Rostedt 374d2d45c7aSSteven Rostedt static int 375d2d45c7aSSteven Rostedt stack_trace_filter_open(struct inode *inode, struct file *file) 376d2d45c7aSSteven Rostedt { 377d2d45c7aSSteven Rostedt return ftrace_regex_open(&trace_ops, FTRACE_ITER_FILTER, 378d2d45c7aSSteven Rostedt inode, file); 379d2d45c7aSSteven Rostedt } 380d2d45c7aSSteven Rostedt 381d2d45c7aSSteven Rostedt static const struct file_operations stack_trace_filter_fops = { 382d2d45c7aSSteven Rostedt .open = stack_trace_filter_open, 383d2d45c7aSSteven Rostedt .read = seq_read, 384d2d45c7aSSteven Rostedt .write = ftrace_filter_write, 385*098c879eSSteven Rostedt (Red Hat) .llseek = tracing_lseek, 386d2d45c7aSSteven Rostedt .release = ftrace_regex_release, 387d2d45c7aSSteven Rostedt }; 388d2d45c7aSSteven Rostedt 389f38f1d2aSSteven Rostedt int 390f38f1d2aSSteven Rostedt stack_trace_sysctl(struct ctl_table *table, int write, 3918d65af78SAlexey Dobriyan void __user *buffer, size_t *lenp, 392f38f1d2aSSteven Rostedt loff_t *ppos) 393f38f1d2aSSteven Rostedt { 394f38f1d2aSSteven Rostedt int ret; 395f38f1d2aSSteven Rostedt 396f38f1d2aSSteven Rostedt mutex_lock(&stack_sysctl_mutex); 397f38f1d2aSSteven Rostedt 3988d65af78SAlexey Dobriyan ret = proc_dointvec(table, write, buffer, lenp, ppos); 399f38f1d2aSSteven Rostedt 400f38f1d2aSSteven Rostedt if (ret || !write || 401a32c7765SLi Zefan (last_stack_tracer_enabled == !!stack_tracer_enabled)) 402f38f1d2aSSteven Rostedt goto out; 403f38f1d2aSSteven Rostedt 404a32c7765SLi Zefan last_stack_tracer_enabled = !!stack_tracer_enabled; 405f38f1d2aSSteven Rostedt 406f38f1d2aSSteven Rostedt if (stack_tracer_enabled) 407f38f1d2aSSteven Rostedt register_ftrace_function(&trace_ops); 408f38f1d2aSSteven Rostedt else 409f38f1d2aSSteven Rostedt unregister_ftrace_function(&trace_ops); 410f38f1d2aSSteven Rostedt 411f38f1d2aSSteven Rostedt out: 412f38f1d2aSSteven Rostedt mutex_unlock(&stack_sysctl_mutex); 413f38f1d2aSSteven Rostedt return ret; 414f38f1d2aSSteven Rostedt } 415f38f1d2aSSteven Rostedt 416762e1207SSteven Rostedt static char stack_trace_filter_buf[COMMAND_LINE_SIZE+1] __initdata; 417762e1207SSteven Rostedt 418f38f1d2aSSteven Rostedt static __init int enable_stacktrace(char *str) 419f38f1d2aSSteven Rostedt { 420762e1207SSteven Rostedt if (strncmp(str, "_filter=", 8) == 0) 421762e1207SSteven Rostedt strncpy(stack_trace_filter_buf, str+8, COMMAND_LINE_SIZE); 422762e1207SSteven Rostedt 423e05a43b7SSteven Rostedt stack_tracer_enabled = 1; 424e05a43b7SSteven Rostedt last_stack_tracer_enabled = 1; 425f38f1d2aSSteven Rostedt return 1; 426f38f1d2aSSteven Rostedt } 427f38f1d2aSSteven Rostedt __setup("stacktrace", enable_stacktrace); 428f38f1d2aSSteven Rostedt 429e5a81b62SSteven Rostedt static __init int stack_trace_init(void) 430e5a81b62SSteven Rostedt { 431e5a81b62SSteven Rostedt struct dentry *d_tracer; 432e5a81b62SSteven Rostedt 433e5a81b62SSteven Rostedt d_tracer = tracing_init_dentry(); 434ed6f1c99SNamhyung Kim if (!d_tracer) 435ed6f1c99SNamhyung Kim return 0; 436e5a81b62SSteven Rostedt 4375452af66SFrederic Weisbecker trace_create_file("stack_max_size", 0644, d_tracer, 438e5a81b62SSteven Rostedt &max_stack_size, &stack_max_size_fops); 439e5a81b62SSteven Rostedt 4405452af66SFrederic Weisbecker trace_create_file("stack_trace", 0444, d_tracer, 441e5a81b62SSteven Rostedt NULL, &stack_trace_fops); 442e5a81b62SSteven Rostedt 443d2d45c7aSSteven Rostedt trace_create_file("stack_trace_filter", 0444, d_tracer, 444d2d45c7aSSteven Rostedt NULL, &stack_trace_filter_fops); 445d2d45c7aSSteven Rostedt 446762e1207SSteven Rostedt if (stack_trace_filter_buf[0]) 447762e1207SSteven Rostedt ftrace_set_early_filter(&trace_ops, stack_trace_filter_buf, 1); 448762e1207SSteven Rostedt 449e05a43b7SSteven Rostedt if (stack_tracer_enabled) 450e5a81b62SSteven Rostedt register_ftrace_function(&trace_ops); 451e5a81b62SSteven Rostedt 452e5a81b62SSteven Rostedt return 0; 453e5a81b62SSteven Rostedt } 454e5a81b62SSteven Rostedt 455e5a81b62SSteven Rostedt device_initcall(stack_trace_init); 456