2 * local apic based NMI watchdog for various CPUs.
4 * This file also handles reservation of performance counters for coordination
5 * with other users (like oprofile).
7 * Note that these events normally don't tick when the CPU idles. This means
8 * the frequency varies with CPU load.
10 * Original code for K7/P6 written by Keith Owens
14 #include <linux/percpu.h>
15 #include <linux/module.h>
16 #include <linux/kernel.h>
17 #include <linux/bitops.h>
18 #include <linux/smp.h>
19 #include <linux/nmi.h>
20 #include <linux/kprobes.h>
23 #include <asm/perf_event.h>
25 struct nmi_watchdog_ctlblk
{
26 unsigned int cccr_msr
;
27 unsigned int perfctr_msr
; /* the MSR to reset in NMI handler */
28 unsigned int evntsel_msr
; /* the MSR to select the events to handle */
31 /* Interface defining a CPU specific perfctr watchdog */
34 void (*unreserve
)(void);
35 int (*setup
)(unsigned nmi_hz
);
36 void (*rearm
)(struct nmi_watchdog_ctlblk
*wd
, unsigned nmi_hz
);
43 static const struct wd_ops
*wd_ops
;
46 * this number is calculated from Intel's MSR_P4_CRU_ESCR5 register and it's
47 * offset from MSR_P4_BSU_ESCR0.
49 * It will be the max for all platforms (for now)
51 #define NMI_MAX_COUNTER_BITS 66
54 * perfctr_nmi_owner tracks the ownership of the perfctr registers:
55 * evtsel_nmi_owner tracks the ownership of the event selection
56 * - different performance counters/ event selection may be reserved for
57 * different subsystems this reservation system just tries to coordinate
60 static DECLARE_BITMAP(perfctr_nmi_owner
, NMI_MAX_COUNTER_BITS
);
61 static DECLARE_BITMAP(evntsel_nmi_owner
, NMI_MAX_COUNTER_BITS
);
63 static DEFINE_PER_CPU(struct nmi_watchdog_ctlblk
, nmi_watchdog_ctlblk
);
65 /* converts an msr to an appropriate reservation bit */
66 static inline unsigned int nmi_perfctr_msr_to_bit(unsigned int msr
)
68 /* returns the bit offset of the performance counter register */
69 switch (boot_cpu_data
.x86_vendor
) {
71 return msr
- MSR_K7_PERFCTR0
;
72 case X86_VENDOR_INTEL
:
73 if (cpu_has(&boot_cpu_data
, X86_FEATURE_ARCH_PERFMON
))
74 return msr
- MSR_ARCH_PERFMON_PERFCTR0
;
76 switch (boot_cpu_data
.x86
) {
78 return msr
- MSR_P6_PERFCTR0
;
80 return msr
- MSR_P4_BPU_PERFCTR0
;
87 * converts an msr to an appropriate reservation bit
88 * returns the bit offset of the event selection register
90 static inline unsigned int nmi_evntsel_msr_to_bit(unsigned int msr
)
92 /* returns the bit offset of the event selection register */
93 switch (boot_cpu_data
.x86_vendor
) {
95 return msr
- MSR_K7_EVNTSEL0
;
96 case X86_VENDOR_INTEL
:
97 if (cpu_has(&boot_cpu_data
, X86_FEATURE_ARCH_PERFMON
))
98 return msr
- MSR_ARCH_PERFMON_EVENTSEL0
;
100 switch (boot_cpu_data
.x86
) {
102 return msr
- MSR_P6_EVNTSEL0
;
104 return msr
- MSR_P4_BSU_ESCR0
;
111 /* checks for a bit availability (hack for oprofile) */
112 int avail_to_resrv_perfctr_nmi_bit(unsigned int counter
)
114 BUG_ON(counter
> NMI_MAX_COUNTER_BITS
);
116 return !test_bit(counter
, perfctr_nmi_owner
);
119 /* checks the an msr for availability */
120 int avail_to_resrv_perfctr_nmi(unsigned int msr
)
122 unsigned int counter
;
124 counter
= nmi_perfctr_msr_to_bit(msr
);
125 BUG_ON(counter
> NMI_MAX_COUNTER_BITS
);
127 return !test_bit(counter
, perfctr_nmi_owner
);
129 EXPORT_SYMBOL(avail_to_resrv_perfctr_nmi_bit
);
131 int reserve_perfctr_nmi(unsigned int msr
)
133 unsigned int counter
;
135 counter
= nmi_perfctr_msr_to_bit(msr
);
136 /* register not managed by the allocator? */
137 if (counter
> NMI_MAX_COUNTER_BITS
)
140 if (!test_and_set_bit(counter
, perfctr_nmi_owner
))
144 EXPORT_SYMBOL(reserve_perfctr_nmi
);
146 void release_perfctr_nmi(unsigned int msr
)
148 unsigned int counter
;
150 counter
= nmi_perfctr_msr_to_bit(msr
);
151 /* register not managed by the allocator? */
152 if (counter
> NMI_MAX_COUNTER_BITS
)
155 clear_bit(counter
, perfctr_nmi_owner
);
157 EXPORT_SYMBOL(release_perfctr_nmi
);
159 int reserve_evntsel_nmi(unsigned int msr
)
161 unsigned int counter
;
163 counter
= nmi_evntsel_msr_to_bit(msr
);
164 /* register not managed by the allocator? */
165 if (counter
> NMI_MAX_COUNTER_BITS
)
168 if (!test_and_set_bit(counter
, evntsel_nmi_owner
))
172 EXPORT_SYMBOL(reserve_evntsel_nmi
);
174 void release_evntsel_nmi(unsigned int msr
)
176 unsigned int counter
;
178 counter
= nmi_evntsel_msr_to_bit(msr
);
179 /* register not managed by the allocator? */
180 if (counter
> NMI_MAX_COUNTER_BITS
)
183 clear_bit(counter
, evntsel_nmi_owner
);
185 EXPORT_SYMBOL(release_evntsel_nmi
);
187 void disable_lapic_nmi_watchdog(void)
189 BUG_ON(nmi_watchdog
!= NMI_LOCAL_APIC
);
191 if (atomic_read(&nmi_active
) <= 0)
194 on_each_cpu(stop_apic_nmi_watchdog
, NULL
, 1);
199 BUG_ON(atomic_read(&nmi_active
) != 0);
202 void enable_lapic_nmi_watchdog(void)
204 BUG_ON(nmi_watchdog
!= NMI_LOCAL_APIC
);
206 /* are we already enabled */
207 if (atomic_read(&nmi_active
) != 0)
210 /* are we lapic aware */
213 if (!wd_ops
->reserve()) {
214 printk(KERN_ERR
"NMI watchdog: cannot reserve perfctrs\n");
218 on_each_cpu(setup_apic_nmi_watchdog
, NULL
, 1);
219 touch_nmi_watchdog();
223 * Activate the NMI watchdog via the local APIC.
226 static unsigned int adjust_for_32bit_ctr(unsigned int hz
)
229 unsigned int retval
= hz
;
232 * On Intel CPUs with P6/ARCH_PERFMON only 32 bits in the counter
233 * are writable, with higher bits sign extending from bit 31.
234 * So, we can only program the counter with 31 bit values and
235 * 32nd bit should be 1, for 33.. to be 1.
236 * Find the appropriate nmi_hz
238 counter_val
= (u64
)cpu_khz
* 1000;
239 do_div(counter_val
, retval
);
240 if (counter_val
> 0x7fffffffULL
) {
241 u64 count
= (u64
)cpu_khz
* 1000;
242 do_div(count
, 0x7fffffffUL
);
248 static void write_watchdog_counter(unsigned int perfctr_msr
,
249 const char *descr
, unsigned nmi_hz
)
251 u64 count
= (u64
)cpu_khz
* 1000;
253 do_div(count
, nmi_hz
);
255 pr_debug("setting %s to -0x%08Lx\n", descr
, count
);
256 wrmsrl(perfctr_msr
, 0 - count
);
259 static void write_watchdog_counter32(unsigned int perfctr_msr
,
260 const char *descr
, unsigned nmi_hz
)
262 u64 count
= (u64
)cpu_khz
* 1000;
264 do_div(count
, nmi_hz
);
266 pr_debug("setting %s to -0x%08Lx\n", descr
, count
);
267 wrmsr(perfctr_msr
, (u32
)(-count
), 0);
271 * AMD K7/K8/Family10h/Family11h support.
272 * AMD keeps this interface nicely stable so there is not much variety
274 #define K7_EVNTSEL_ENABLE (1 << 22)
275 #define K7_EVNTSEL_INT (1 << 20)
276 #define K7_EVNTSEL_OS (1 << 17)
277 #define K7_EVNTSEL_USR (1 << 16)
278 #define K7_EVENT_CYCLES_PROCESSOR_IS_RUNNING 0x76
279 #define K7_NMI_EVENT K7_EVENT_CYCLES_PROCESSOR_IS_RUNNING
281 static int setup_k7_watchdog(unsigned nmi_hz
)
283 unsigned int perfctr_msr
, evntsel_msr
;
284 unsigned int evntsel
;
285 struct nmi_watchdog_ctlblk
*wd
= &__get_cpu_var(nmi_watchdog_ctlblk
);
287 perfctr_msr
= wd_ops
->perfctr
;
288 evntsel_msr
= wd_ops
->evntsel
;
290 wrmsrl(perfctr_msr
, 0UL);
292 evntsel
= K7_EVNTSEL_INT
297 /* setup the timer */
298 wrmsr(evntsel_msr
, evntsel
, 0);
299 write_watchdog_counter(perfctr_msr
, "K7_PERFCTR0", nmi_hz
);
301 /* initialize the wd struct before enabling */
302 wd
->perfctr_msr
= perfctr_msr
;
303 wd
->evntsel_msr
= evntsel_msr
;
304 wd
->cccr_msr
= 0; /* unused */
306 /* ok, everything is initialized, announce that we're set */
307 cpu_nmi_set_wd_enabled();
309 apic_write(APIC_LVTPC
, APIC_DM_NMI
);
310 evntsel
|= K7_EVNTSEL_ENABLE
;
311 wrmsr(evntsel_msr
, evntsel
, 0);
316 static void single_msr_stop_watchdog(void)
318 struct nmi_watchdog_ctlblk
*wd
= &__get_cpu_var(nmi_watchdog_ctlblk
);
320 wrmsr(wd
->evntsel_msr
, 0, 0);
323 static int single_msr_reserve(void)
325 if (!reserve_perfctr_nmi(wd_ops
->perfctr
))
328 if (!reserve_evntsel_nmi(wd_ops
->evntsel
)) {
329 release_perfctr_nmi(wd_ops
->perfctr
);
335 static void single_msr_unreserve(void)
337 release_evntsel_nmi(wd_ops
->evntsel
);
338 release_perfctr_nmi(wd_ops
->perfctr
);
341 static void __kprobes
342 single_msr_rearm(struct nmi_watchdog_ctlblk
*wd
, unsigned nmi_hz
)
344 /* start the cycle over again */
345 write_watchdog_counter(wd
->perfctr_msr
, NULL
, nmi_hz
);
348 static const struct wd_ops k7_wd_ops
= {
349 .reserve
= single_msr_reserve
,
350 .unreserve
= single_msr_unreserve
,
351 .setup
= setup_k7_watchdog
,
352 .rearm
= single_msr_rearm
,
353 .stop
= single_msr_stop_watchdog
,
354 .perfctr
= MSR_K7_PERFCTR0
,
355 .evntsel
= MSR_K7_EVNTSEL0
,
356 .checkbit
= 1ULL << 47,
360 * Intel Model 6 (PPro+,P2,P3,P-M,Core1)
362 #define P6_EVNTSEL0_ENABLE (1 << 22)
363 #define P6_EVNTSEL_INT (1 << 20)
364 #define P6_EVNTSEL_OS (1 << 17)
365 #define P6_EVNTSEL_USR (1 << 16)
366 #define P6_EVENT_CPU_CLOCKS_NOT_HALTED 0x79
367 #define P6_NMI_EVENT P6_EVENT_CPU_CLOCKS_NOT_HALTED
369 static int setup_p6_watchdog(unsigned nmi_hz
)
371 unsigned int perfctr_msr
, evntsel_msr
;
372 unsigned int evntsel
;
373 struct nmi_watchdog_ctlblk
*wd
= &__get_cpu_var(nmi_watchdog_ctlblk
);
375 perfctr_msr
= wd_ops
->perfctr
;
376 evntsel_msr
= wd_ops
->evntsel
;
378 /* KVM doesn't implement this MSR */
379 if (wrmsr_safe(perfctr_msr
, 0, 0) < 0)
382 evntsel
= P6_EVNTSEL_INT
387 /* setup the timer */
388 wrmsr(evntsel_msr
, evntsel
, 0);
389 nmi_hz
= adjust_for_32bit_ctr(nmi_hz
);
390 write_watchdog_counter32(perfctr_msr
, "P6_PERFCTR0", nmi_hz
);
392 /* initialize the wd struct before enabling */
393 wd
->perfctr_msr
= perfctr_msr
;
394 wd
->evntsel_msr
= evntsel_msr
;
395 wd
->cccr_msr
= 0; /* unused */
397 /* ok, everything is initialized, announce that we're set */
398 cpu_nmi_set_wd_enabled();
400 apic_write(APIC_LVTPC
, APIC_DM_NMI
);
401 evntsel
|= P6_EVNTSEL0_ENABLE
;
402 wrmsr(evntsel_msr
, evntsel
, 0);
407 static void __kprobes
p6_rearm(struct nmi_watchdog_ctlblk
*wd
, unsigned nmi_hz
)
410 * P6 based Pentium M need to re-unmask
411 * the apic vector but it doesn't hurt
413 * ArchPerfom/Core Duo also needs this
415 apic_write(APIC_LVTPC
, APIC_DM_NMI
);
417 /* P6/ARCH_PERFMON has 32 bit counter write */
418 write_watchdog_counter32(wd
->perfctr_msr
, NULL
, nmi_hz
);
421 static const struct wd_ops p6_wd_ops
= {
422 .reserve
= single_msr_reserve
,
423 .unreserve
= single_msr_unreserve
,
424 .setup
= setup_p6_watchdog
,
426 .stop
= single_msr_stop_watchdog
,
427 .perfctr
= MSR_P6_PERFCTR0
,
428 .evntsel
= MSR_P6_EVNTSEL0
,
429 .checkbit
= 1ULL << 39,
433 * Intel P4 performance counters.
434 * By far the most complicated of all.
436 #define MSR_P4_MISC_ENABLE_PERF_AVAIL (1 << 7)
437 #define P4_ESCR_EVENT_SELECT(N) ((N) << 25)
438 #define P4_ESCR_OS (1 << 3)
439 #define P4_ESCR_USR (1 << 2)
440 #define P4_CCCR_OVF_PMI0 (1 << 26)
441 #define P4_CCCR_OVF_PMI1 (1 << 27)
442 #define P4_CCCR_THRESHOLD(N) ((N) << 20)
443 #define P4_CCCR_COMPLEMENT (1 << 19)
444 #define P4_CCCR_COMPARE (1 << 18)
445 #define P4_CCCR_REQUIRED (3 << 16)
446 #define P4_CCCR_ESCR_SELECT(N) ((N) << 13)
447 #define P4_CCCR_ENABLE (1 << 12)
448 #define P4_CCCR_OVF (1 << 31)
450 #define P4_CONTROLS 18
451 static unsigned int p4_controls
[18] = {
472 * Set up IQ_COUNTER0 to behave like a clock, by having IQ_CCCR0 filter
473 * CRU_ESCR0 (with any non-null event selector) through a complemented
474 * max threshold. [IA32-Vol3, Section 14.9.9]
476 static int setup_p4_watchdog(unsigned nmi_hz
)
478 unsigned int perfctr_msr
, evntsel_msr
, cccr_msr
;
479 unsigned int evntsel
, cccr_val
;
480 unsigned int misc_enable
, dummy
;
482 struct nmi_watchdog_ctlblk
*wd
= &__get_cpu_var(nmi_watchdog_ctlblk
);
484 rdmsr(MSR_IA32_MISC_ENABLE
, misc_enable
, dummy
);
485 if (!(misc_enable
& MSR_P4_MISC_ENABLE_PERF_AVAIL
))
489 /* detect which hyperthread we are on */
490 if (smp_num_siblings
== 2) {
491 unsigned int ebx
, apicid
;
494 apicid
= (ebx
>> 24) & 0xff;
501 * performance counters are shared resources
502 * assign each hyperthread its own set
503 * (re-use the ESCR0 register, seems safe
504 * and keeps the cccr_val the same)
508 perfctr_msr
= MSR_P4_IQ_PERFCTR0
;
509 evntsel_msr
= MSR_P4_CRU_ESCR0
;
510 cccr_msr
= MSR_P4_IQ_CCCR0
;
511 cccr_val
= P4_CCCR_OVF_PMI0
| P4_CCCR_ESCR_SELECT(4);
514 * If we're on the kdump kernel or other situation, we may
515 * still have other performance counter registers set to
516 * interrupt and they'll keep interrupting forever because
517 * of the P4_CCCR_OVF quirk. So we need to ACK all the
518 * pending interrupts and disable all the registers here,
519 * before reenabling the NMI delivery. Refer to p4_rearm()
520 * about the P4_CCCR_OVF quirk.
523 unsigned int low
, high
;
526 for (i
= 0; i
< P4_CONTROLS
; i
++) {
527 rdmsr(p4_controls
[i
], low
, high
);
528 low
&= ~(P4_CCCR_ENABLE
| P4_CCCR_OVF
);
529 wrmsr(p4_controls
[i
], low
, high
);
534 perfctr_msr
= MSR_P4_IQ_PERFCTR1
;
535 evntsel_msr
= MSR_P4_CRU_ESCR0
;
536 cccr_msr
= MSR_P4_IQ_CCCR1
;
538 /* Pentium 4 D processors don't support P4_CCCR_OVF_PMI1 */
539 if (boot_cpu_data
.x86_model
== 4 && boot_cpu_data
.x86_mask
== 4)
540 cccr_val
= P4_CCCR_OVF_PMI0
;
542 cccr_val
= P4_CCCR_OVF_PMI1
;
543 cccr_val
|= P4_CCCR_ESCR_SELECT(4);
546 evntsel
= P4_ESCR_EVENT_SELECT(0x3F)
550 cccr_val
|= P4_CCCR_THRESHOLD(15)
555 wrmsr(evntsel_msr
, evntsel
, 0);
556 wrmsr(cccr_msr
, cccr_val
, 0);
557 write_watchdog_counter(perfctr_msr
, "P4_IQ_COUNTER0", nmi_hz
);
559 wd
->perfctr_msr
= perfctr_msr
;
560 wd
->evntsel_msr
= evntsel_msr
;
561 wd
->cccr_msr
= cccr_msr
;
563 /* ok, everything is initialized, announce that we're set */
564 cpu_nmi_set_wd_enabled();
566 apic_write(APIC_LVTPC
, APIC_DM_NMI
);
567 cccr_val
|= P4_CCCR_ENABLE
;
568 wrmsr(cccr_msr
, cccr_val
, 0);
572 static void stop_p4_watchdog(void)
574 struct nmi_watchdog_ctlblk
*wd
= &__get_cpu_var(nmi_watchdog_ctlblk
);
575 wrmsr(wd
->cccr_msr
, 0, 0);
576 wrmsr(wd
->evntsel_msr
, 0, 0);
579 static int p4_reserve(void)
581 if (!reserve_perfctr_nmi(MSR_P4_IQ_PERFCTR0
))
584 if (smp_num_siblings
> 1 && !reserve_perfctr_nmi(MSR_P4_IQ_PERFCTR1
))
587 if (!reserve_evntsel_nmi(MSR_P4_CRU_ESCR0
))
589 /* RED-PEN why is ESCR1 not reserved here? */
593 if (smp_num_siblings
> 1)
594 release_perfctr_nmi(MSR_P4_IQ_PERFCTR1
);
597 release_perfctr_nmi(MSR_P4_IQ_PERFCTR0
);
601 static void p4_unreserve(void)
604 if (smp_num_siblings
> 1)
605 release_perfctr_nmi(MSR_P4_IQ_PERFCTR1
);
607 release_evntsel_nmi(MSR_P4_CRU_ESCR0
);
608 release_perfctr_nmi(MSR_P4_IQ_PERFCTR0
);
611 static void __kprobes
p4_rearm(struct nmi_watchdog_ctlblk
*wd
, unsigned nmi_hz
)
616 * - An overflown perfctr will assert its interrupt
617 * until the OVF flag in its CCCR is cleared.
618 * - LVTPC is masked on interrupt and must be
619 * unmasked by the LVTPC handler.
621 rdmsrl(wd
->cccr_msr
, dummy
);
622 dummy
&= ~P4_CCCR_OVF
;
623 wrmsrl(wd
->cccr_msr
, dummy
);
624 apic_write(APIC_LVTPC
, APIC_DM_NMI
);
625 /* start the cycle over again */
626 write_watchdog_counter(wd
->perfctr_msr
, NULL
, nmi_hz
);
629 static const struct wd_ops p4_wd_ops
= {
630 .reserve
= p4_reserve
,
631 .unreserve
= p4_unreserve
,
632 .setup
= setup_p4_watchdog
,
634 .stop
= stop_p4_watchdog
,
635 /* RED-PEN this is wrong for the other sibling */
636 .perfctr
= MSR_P4_BPU_PERFCTR0
,
637 .evntsel
= MSR_P4_BSU_ESCR0
,
638 .checkbit
= 1ULL << 39,
642 * Watchdog using the Intel architected PerfMon.
643 * Used for Core2 and hopefully all future Intel CPUs.
645 #define ARCH_PERFMON_NMI_EVENT_SEL ARCH_PERFMON_UNHALTED_CORE_CYCLES_SEL
646 #define ARCH_PERFMON_NMI_EVENT_UMASK ARCH_PERFMON_UNHALTED_CORE_CYCLES_UMASK
648 static struct wd_ops intel_arch_wd_ops
;
650 static int setup_intel_arch_watchdog(unsigned nmi_hz
)
653 union cpuid10_eax eax
;
655 unsigned int perfctr_msr
, evntsel_msr
;
656 unsigned int evntsel
;
657 struct nmi_watchdog_ctlblk
*wd
= &__get_cpu_var(nmi_watchdog_ctlblk
);
660 * Check whether the Architectural PerfMon supports
661 * Unhalted Core Cycles Event or not.
662 * NOTE: Corresponding bit = 0 in ebx indicates event present.
664 cpuid(10, &(eax
.full
), &ebx
, &unused
, &unused
);
665 if ((eax
.split
.mask_length
<
666 (ARCH_PERFMON_UNHALTED_CORE_CYCLES_INDEX
+1)) ||
667 (ebx
& ARCH_PERFMON_UNHALTED_CORE_CYCLES_PRESENT
))
670 perfctr_msr
= wd_ops
->perfctr
;
671 evntsel_msr
= wd_ops
->evntsel
;
673 wrmsrl(perfctr_msr
, 0UL);
675 evntsel
= ARCH_PERFMON_EVENTSEL_INT
676 | ARCH_PERFMON_EVENTSEL_OS
677 | ARCH_PERFMON_EVENTSEL_USR
678 | ARCH_PERFMON_NMI_EVENT_SEL
679 | ARCH_PERFMON_NMI_EVENT_UMASK
;
681 /* setup the timer */
682 wrmsr(evntsel_msr
, evntsel
, 0);
683 nmi_hz
= adjust_for_32bit_ctr(nmi_hz
);
684 write_watchdog_counter32(perfctr_msr
, "INTEL_ARCH_PERFCTR0", nmi_hz
);
686 wd
->perfctr_msr
= perfctr_msr
;
687 wd
->evntsel_msr
= evntsel_msr
;
688 wd
->cccr_msr
= 0; /* unused */
690 /* ok, everything is initialized, announce that we're set */
691 cpu_nmi_set_wd_enabled();
693 apic_write(APIC_LVTPC
, APIC_DM_NMI
);
694 evntsel
|= ARCH_PERFMON_EVENTSEL0_ENABLE
;
695 wrmsr(evntsel_msr
, evntsel
, 0);
696 intel_arch_wd_ops
.checkbit
= 1ULL << (eax
.split
.bit_width
- 1);
700 static struct wd_ops intel_arch_wd_ops __read_mostly
= {
701 .reserve
= single_msr_reserve
,
702 .unreserve
= single_msr_unreserve
,
703 .setup
= setup_intel_arch_watchdog
,
705 .stop
= single_msr_stop_watchdog
,
706 .perfctr
= MSR_ARCH_PERFMON_PERFCTR1
,
707 .evntsel
= MSR_ARCH_PERFMON_EVENTSEL1
,
710 static void probe_nmi_watchdog(void)
712 switch (boot_cpu_data
.x86_vendor
) {
714 if (boot_cpu_data
.x86
!= 6 && boot_cpu_data
.x86
!= 15 &&
715 boot_cpu_data
.x86
!= 16)
719 case X86_VENDOR_INTEL
:
720 /* Work around where perfctr1 doesn't have a working enable
721 * bit as described in the following errata:
722 * AE49 Core Duo and Intel Core Solo 65 nm
723 * AN49 Intel Pentium Dual-Core
724 * AF49 Dual-Core Intel Xeon Processor LV
726 if ((boot_cpu_data
.x86
== 6 && boot_cpu_data
.x86_model
== 14) ||
727 ((boot_cpu_data
.x86
== 6 && boot_cpu_data
.x86_model
== 15 &&
728 boot_cpu_data
.x86_mask
== 4))) {
729 intel_arch_wd_ops
.perfctr
= MSR_ARCH_PERFMON_PERFCTR0
;
730 intel_arch_wd_ops
.evntsel
= MSR_ARCH_PERFMON_EVENTSEL0
;
732 if (cpu_has(&boot_cpu_data
, X86_FEATURE_ARCH_PERFMON
)) {
733 wd_ops
= &intel_arch_wd_ops
;
736 switch (boot_cpu_data
.x86
) {
738 if (boot_cpu_data
.x86_model
> 13)
753 /* Interface to nmi.c */
755 int lapic_watchdog_init(unsigned nmi_hz
)
758 probe_nmi_watchdog();
760 printk(KERN_INFO
"NMI watchdog: CPU not supported\n");
764 if (!wd_ops
->reserve()) {
766 "NMI watchdog: cannot reserve perfctrs\n");
771 if (!(wd_ops
->setup(nmi_hz
))) {
772 printk(KERN_ERR
"Cannot setup NMI watchdog on CPU %d\n",
773 raw_smp_processor_id());
780 void lapic_watchdog_stop(void)
786 unsigned lapic_adjust_nmi_hz(unsigned hz
)
788 struct nmi_watchdog_ctlblk
*wd
= &__get_cpu_var(nmi_watchdog_ctlblk
);
789 if (wd
->perfctr_msr
== MSR_P6_PERFCTR0
||
790 wd
->perfctr_msr
== MSR_ARCH_PERFMON_PERFCTR1
)
791 hz
= adjust_for_32bit_ctr(hz
);
795 int __kprobes
lapic_wd_event(unsigned nmi_hz
)
797 struct nmi_watchdog_ctlblk
*wd
= &__get_cpu_var(nmi_watchdog_ctlblk
);
800 rdmsrl(wd
->perfctr_msr
, ctr
);
801 if (ctr
& wd_ops
->checkbit
) /* perfctr still running? */
804 wd_ops
->rearm(wd
, nmi_hz
);