Staging: hv: mousevsc: Get rid of the unused state: num_outstanding_req
[zen-stable.git] / arch / ia64 / kernel / crash.c
blobb942f4032d7a537856e86d9fc69e83b4fe16dcd3
1 /*
2 * arch/ia64/kernel/crash.c
4 * Architecture specific (ia64) functions for kexec based crash dumps.
6 * Created by: Khalid Aziz <khalid.aziz@hp.com>
7 * Copyright (C) 2005 Hewlett-Packard Development Company, L.P.
8 * Copyright (C) 2005 Intel Corp Zou Nan hai <nanhai.zou@intel.com>
11 #include <linux/smp.h>
12 #include <linux/delay.h>
13 #include <linux/crash_dump.h>
14 #include <linux/bootmem.h>
15 #include <linux/kexec.h>
16 #include <linux/elfcore.h>
17 #include <linux/sysctl.h>
18 #include <linux/init.h>
19 #include <linux/kdebug.h>
21 #include <asm/mca.h>
23 int kdump_status[NR_CPUS];
24 static atomic_t kdump_cpu_frozen;
25 atomic_t kdump_in_progress;
26 static int kdump_freeze_monarch;
27 static int kdump_on_init = 1;
28 static int kdump_on_fatal_mca = 1;
30 static inline Elf64_Word
31 *append_elf_note(Elf64_Word *buf, char *name, unsigned type, void *data,
32 size_t data_len)
34 struct elf_note *note = (struct elf_note *)buf;
35 note->n_namesz = strlen(name) + 1;
36 note->n_descsz = data_len;
37 note->n_type = type;
38 buf += (sizeof(*note) + 3)/4;
39 memcpy(buf, name, note->n_namesz);
40 buf += (note->n_namesz + 3)/4;
41 memcpy(buf, data, data_len);
42 buf += (data_len + 3)/4;
43 return buf;
46 static void
47 final_note(void *buf)
49 memset(buf, 0, sizeof(struct elf_note));
52 extern void ia64_dump_cpu_regs(void *);
54 static DEFINE_PER_CPU(struct elf_prstatus, elf_prstatus);
56 void
57 crash_save_this_cpu(void)
59 void *buf;
60 unsigned long cfm, sof, sol;
62 int cpu = smp_processor_id();
63 struct elf_prstatus *prstatus = &per_cpu(elf_prstatus, cpu);
65 elf_greg_t *dst = (elf_greg_t *)&(prstatus->pr_reg);
66 memset(prstatus, 0, sizeof(*prstatus));
67 prstatus->pr_pid = current->pid;
69 ia64_dump_cpu_regs(dst);
70 cfm = dst[43];
71 sol = (cfm >> 7) & 0x7f;
72 sof = cfm & 0x7f;
73 dst[46] = (unsigned long)ia64_rse_skip_regs((unsigned long *)dst[46],
74 sof - sol);
76 buf = (u64 *) per_cpu_ptr(crash_notes, cpu);
77 if (!buf)
78 return;
79 buf = append_elf_note(buf, KEXEC_CORE_NOTE_NAME, NT_PRSTATUS, prstatus,
80 sizeof(*prstatus));
81 final_note(buf);
84 #ifdef CONFIG_SMP
85 static int
86 kdump_wait_cpu_freeze(void)
88 int cpu_num = num_online_cpus() - 1;
89 int timeout = 1000;
90 while(timeout-- > 0) {
91 if (atomic_read(&kdump_cpu_frozen) == cpu_num)
92 return 0;
93 udelay(1000);
95 return 1;
97 #endif
99 void
100 machine_crash_shutdown(struct pt_regs *pt)
102 /* This function is only called after the system
103 * has paniced or is otherwise in a critical state.
104 * The minimum amount of code to allow a kexec'd kernel
105 * to run successfully needs to happen here.
107 * In practice this means shooting down the other cpus in
108 * an SMP system.
110 kexec_disable_iosapic();
111 #ifdef CONFIG_SMP
113 * If kdump_on_init is set and an INIT is asserted here, kdump will
114 * be started again via INIT monarch.
116 local_irq_disable();
117 ia64_set_psr_mc(); /* mask MCA/INIT */
118 if (atomic_inc_return(&kdump_in_progress) != 1)
119 unw_init_running(kdump_cpu_freeze, NULL);
122 * Now this cpu is ready for kdump.
123 * Stop all others by IPI or INIT. They could receive INIT from
124 * outside and might be INIT monarch, but only thing they have to
125 * do is falling into kdump_cpu_freeze().
127 * If an INIT is asserted here:
128 * - All receivers might be slaves, since some of cpus could already
129 * be frozen and INIT might be masked on monarch. In this case,
130 * all slaves will be frozen soon since kdump_in_progress will let
131 * them into DIE_INIT_SLAVE_LEAVE.
132 * - One might be a monarch, but INIT rendezvous will fail since
133 * at least this cpu already have INIT masked so it never join
134 * to the rendezvous. In this case, all slaves and monarch will
135 * be frozen soon with no wait since the INIT rendezvous is skipped
136 * by kdump_in_progress.
138 kdump_smp_send_stop();
139 /* not all cpu response to IPI, send INIT to freeze them */
140 if (kdump_wait_cpu_freeze()) {
141 kdump_smp_send_init();
142 /* wait again, don't go ahead if possible */
143 kdump_wait_cpu_freeze();
145 #endif
148 static void
149 machine_kdump_on_init(void)
151 crash_save_vmcoreinfo();
152 local_irq_disable();
153 kexec_disable_iosapic();
154 machine_kexec(ia64_kimage);
157 void
158 kdump_cpu_freeze(struct unw_frame_info *info, void *arg)
160 int cpuid;
162 local_irq_disable();
163 cpuid = smp_processor_id();
164 crash_save_this_cpu();
165 current->thread.ksp = (__u64)info->sw - 16;
167 ia64_set_psr_mc(); /* mask MCA/INIT and stop reentrance */
169 atomic_inc(&kdump_cpu_frozen);
170 kdump_status[cpuid] = 1;
171 mb();
172 for (;;)
173 cpu_relax();
176 static int
177 kdump_init_notifier(struct notifier_block *self, unsigned long val, void *data)
179 struct ia64_mca_notify_die *nd;
180 struct die_args *args = data;
182 if (atomic_read(&kdump_in_progress)) {
183 switch (val) {
184 case DIE_INIT_MONARCH_LEAVE:
185 if (!kdump_freeze_monarch)
186 break;
187 /* fall through */
188 case DIE_INIT_SLAVE_LEAVE:
189 case DIE_INIT_MONARCH_ENTER:
190 case DIE_MCA_RENDZVOUS_LEAVE:
191 unw_init_running(kdump_cpu_freeze, NULL);
192 break;
196 if (!kdump_on_init && !kdump_on_fatal_mca)
197 return NOTIFY_DONE;
199 if (!ia64_kimage) {
200 if (val == DIE_INIT_MONARCH_LEAVE)
201 ia64_mca_printk(KERN_NOTICE
202 "%s: kdump not configured\n",
203 __func__);
204 return NOTIFY_DONE;
207 if (val != DIE_INIT_MONARCH_LEAVE &&
208 val != DIE_INIT_MONARCH_PROCESS &&
209 val != DIE_MCA_MONARCH_LEAVE)
210 return NOTIFY_DONE;
212 nd = (struct ia64_mca_notify_die *)args->err;
214 switch (val) {
215 case DIE_INIT_MONARCH_PROCESS:
216 /* Reason code 1 means machine check rendezvous*/
217 if (kdump_on_init && (nd->sos->rv_rc != 1)) {
218 if (atomic_inc_return(&kdump_in_progress) != 1)
219 kdump_freeze_monarch = 1;
221 break;
222 case DIE_INIT_MONARCH_LEAVE:
223 /* Reason code 1 means machine check rendezvous*/
224 if (kdump_on_init && (nd->sos->rv_rc != 1))
225 machine_kdump_on_init();
226 break;
227 case DIE_MCA_MONARCH_LEAVE:
228 /* *(nd->data) indicate if MCA is recoverable */
229 if (kdump_on_fatal_mca && !(*(nd->data))) {
230 if (atomic_inc_return(&kdump_in_progress) == 1)
231 machine_kdump_on_init();
232 /* We got fatal MCA while kdump!? No way!! */
234 break;
236 return NOTIFY_DONE;
239 #ifdef CONFIG_SYSCTL
240 static ctl_table kdump_ctl_table[] = {
242 .procname = "kdump_on_init",
243 .data = &kdump_on_init,
244 .maxlen = sizeof(int),
245 .mode = 0644,
246 .proc_handler = proc_dointvec,
249 .procname = "kdump_on_fatal_mca",
250 .data = &kdump_on_fatal_mca,
251 .maxlen = sizeof(int),
252 .mode = 0644,
253 .proc_handler = proc_dointvec,
258 static ctl_table sys_table[] = {
260 .procname = "kernel",
261 .mode = 0555,
262 .child = kdump_ctl_table,
266 #endif
268 static int
269 machine_crash_setup(void)
271 /* be notified before default_monarch_init_process */
272 static struct notifier_block kdump_init_notifier_nb = {
273 .notifier_call = kdump_init_notifier,
274 .priority = 1,
276 int ret;
277 if((ret = register_die_notifier(&kdump_init_notifier_nb)) != 0)
278 return ret;
279 #ifdef CONFIG_SYSCTL
280 register_sysctl_table(sys_table);
281 #endif
282 return 0;
285 __initcall(machine_crash_setup);