Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
unwinder.c
Go to the documentation of this file.
1 /*
2  * Copyright (C) 2009 Matt Fleming
3  *
4  * Based, in part, on kernel/time/clocksource.c.
5  *
6  * This file provides arbitration code for stack unwinders.
7  *
8  * Multiple stack unwinders can be available on a system, usually with
9  * the most accurate unwinder being the currently active one.
10  */
11 #include <linux/errno.h>
12 #include <linux/list.h>
13 #include <linux/spinlock.h>
14 #include <linux/module.h>
15 #include <asm/unwinder.h>
16 #include <linux/atomic.h>
17 
18 /*
19  * This is the most basic stack unwinder an architecture can
20  * provide. For architectures without reliable frame pointers, e.g.
21  * RISC CPUs, it can be implemented by looking through the stack for
22  * addresses that lie within the kernel text section.
23  *
24  * Other CPUs, e.g. x86, can use their frame pointer register to
25  * construct more accurate stack traces.
26  */
27 static struct list_head unwinder_list;
28 static struct unwinder stack_reader = {
29  .name = "stack-reader",
30  .dump = stack_reader_dump,
31  .rating = 50,
32  .list = {
33  .next = &unwinder_list,
34  .prev = &unwinder_list,
35  },
36 };
37 
38 /*
39  * "curr_unwinder" points to the stack unwinder currently in use. This
40  * is the unwinder with the highest rating.
41  *
42  * "unwinder_list" is a linked-list of all available unwinders, sorted
43  * by rating.
44  *
45  * All modifications of "curr_unwinder" and "unwinder_list" must be
46  * performed whilst holding "unwinder_lock".
47  */
48 static struct unwinder *curr_unwinder = &stack_reader;
49 
50 static struct list_head unwinder_list = {
51  .next = &stack_reader.list,
52  .prev = &stack_reader.list,
53 };
54 
55 static DEFINE_SPINLOCK(unwinder_lock);
56 
65 static struct unwinder *select_unwinder(void)
66 {
67  struct unwinder *best;
68 
69  if (list_empty(&unwinder_list))
70  return NULL;
71 
72  best = list_entry(unwinder_list.next, struct unwinder, list);
73  if (best == curr_unwinder)
74  return NULL;
75 
76  return best;
77 }
78 
79 /*
80  * Enqueue the stack unwinder sorted by rating.
81  */
82 static int unwinder_enqueue(struct unwinder *ops)
83 {
84  struct list_head *tmp, *entry = &unwinder_list;
85 
86  list_for_each(tmp, &unwinder_list) {
87  struct unwinder *o;
88 
89  o = list_entry(tmp, struct unwinder, list);
90  if (o == ops)
91  return -EBUSY;
92  /* Keep track of the place, where to insert */
93  if (o->rating >= ops->rating)
94  entry = tmp;
95  }
96  list_add(&ops->list, entry);
97 
98  return 0;
99 }
100 
111 {
112  unsigned long flags;
113  int ret;
114 
115  spin_lock_irqsave(&unwinder_lock, flags);
116  ret = unwinder_enqueue(u);
117  if (!ret)
118  curr_unwinder = select_unwinder();
119  spin_unlock_irqrestore(&unwinder_lock, flags);
120 
121  return ret;
122 }
123 
125 
126 /*
127  * Unwind the call stack and pass information to the stacktrace_ops
128  * functions. Also handle the case where we need to switch to a new
129  * stack dumper because the current one faulted unexpectedly.
130  */
131 void unwind_stack(struct task_struct *task, struct pt_regs *regs,
132  unsigned long *sp, const struct stacktrace_ops *ops,
133  void *data)
134 {
135  unsigned long flags;
136 
137  /*
138  * The problem with unwinders with high ratings is that they are
139  * inherently more complicated than the simple ones with lower
140  * ratings. We are therefore more likely to fault in the
141  * complicated ones, e.g. hitting BUG()s. If we fault in the
142  * code for the current stack unwinder we try to downgrade to
143  * one with a lower rating.
144  *
145  * Hopefully this will give us a semi-reliable stacktrace so we
146  * can diagnose why curr_unwinder->dump() faulted.
147  */
148  if (unwinder_faulted) {
149  spin_lock_irqsave(&unwinder_lock, flags);
150 
151  /* Make sure no one beat us to changing the unwinder */
152  if (unwinder_faulted && !list_is_singular(&unwinder_list)) {
153  list_del(&curr_unwinder->list);
154  curr_unwinder = select_unwinder();
155 
156  unwinder_faulted = 0;
157  }
158 
159  spin_unlock_irqrestore(&unwinder_lock, flags);
160  }
161 
162  curr_unwinder->dump(task, regs, sp, ops, data);
163 }
164 EXPORT_SYMBOL_GPL(unwind_stack);