mtd: nand: omap: Fix comment in platform data using wrong Kconfig symbol
[linux/fpc-iii.git] / arch / x86 / kernel / quirks.c
blob8451f38ad3999fcb95d85c7c453de544745d0c4b
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * This file contains work-arounds for x86 and x86_64 platform bugs.
4 */
5 #include <linux/dmi.h>
6 #include <linux/pci.h>
7 #include <linux/irq.h>
9 #include <asm/hpet.h>
10 #include <asm/setup.h>
12 #if defined(CONFIG_X86_IO_APIC) && defined(CONFIG_SMP) && defined(CONFIG_PCI)
14 static void quirk_intel_irqbalance(struct pci_dev *dev)
16 u8 config;
17 u16 word;
19 /* BIOS may enable hardware IRQ balancing for
20 * E7520/E7320/E7525(revision ID 0x9 and below)
21 * based platforms.
22 * Disable SW irqbalance/affinity on those platforms.
24 if (dev->revision > 0x9)
25 return;
27 /* enable access to config space*/
28 pci_read_config_byte(dev, 0xf4, &config);
29 pci_write_config_byte(dev, 0xf4, config|0x2);
32 * read xTPR register. We may not have a pci_dev for device 8
33 * because it might be hidden until the above write.
35 pci_bus_read_config_word(dev->bus, PCI_DEVFN(8, 0), 0x4c, &word);
37 if (!(word & (1 << 13))) {
38 dev_info(&dev->dev, "Intel E7520/7320/7525 detected; "
39 "disabling irq balancing and affinity\n");
40 noirqdebug_setup("");
41 #ifdef CONFIG_PROC_FS
42 no_irq_affinity = 1;
43 #endif
46 /* put back the original value for config space*/
47 if (!(config & 0x2))
48 pci_write_config_byte(dev, 0xf4, config);
50 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7320_MCH,
51 quirk_intel_irqbalance);
52 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7525_MCH,
53 quirk_intel_irqbalance);
54 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7520_MCH,
55 quirk_intel_irqbalance);
56 #endif
58 #if defined(CONFIG_HPET_TIMER)
59 unsigned long force_hpet_address;
61 static enum {
62 NONE_FORCE_HPET_RESUME,
63 OLD_ICH_FORCE_HPET_RESUME,
64 ICH_FORCE_HPET_RESUME,
65 VT8237_FORCE_HPET_RESUME,
66 NVIDIA_FORCE_HPET_RESUME,
67 ATI_FORCE_HPET_RESUME,
68 } force_hpet_resume_type;
70 static void __iomem *rcba_base;
72 static void ich_force_hpet_resume(void)
74 u32 val;
76 if (!force_hpet_address)
77 return;
79 BUG_ON(rcba_base == NULL);
81 /* read the Function Disable register, dword mode only */
82 val = readl(rcba_base + 0x3404);
83 if (!(val & 0x80)) {
84 /* HPET disabled in HPTC. Trying to enable */
85 writel(val | 0x80, rcba_base + 0x3404);
88 val = readl(rcba_base + 0x3404);
89 if (!(val & 0x80))
90 BUG();
91 else
92 printk(KERN_DEBUG "Force enabled HPET at resume\n");
94 return;
97 static void ich_force_enable_hpet(struct pci_dev *dev)
99 u32 val;
100 u32 uninitialized_var(rcba);
101 int err = 0;
103 if (hpet_address || force_hpet_address)
104 return;
106 pci_read_config_dword(dev, 0xF0, &rcba);
107 rcba &= 0xFFFFC000;
108 if (rcba == 0) {
109 dev_printk(KERN_DEBUG, &dev->dev, "RCBA disabled; "
110 "cannot force enable HPET\n");
111 return;
114 /* use bits 31:14, 16 kB aligned */
115 rcba_base = ioremap_nocache(rcba, 0x4000);
116 if (rcba_base == NULL) {
117 dev_printk(KERN_DEBUG, &dev->dev, "ioremap failed; "
118 "cannot force enable HPET\n");
119 return;
122 /* read the Function Disable register, dword mode only */
123 val = readl(rcba_base + 0x3404);
125 if (val & 0x80) {
126 /* HPET is enabled in HPTC. Just not reported by BIOS */
127 val = val & 0x3;
128 force_hpet_address = 0xFED00000 | (val << 12);
129 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
130 "0x%lx\n", force_hpet_address);
131 iounmap(rcba_base);
132 return;
135 /* HPET disabled in HPTC. Trying to enable */
136 writel(val | 0x80, rcba_base + 0x3404);
138 val = readl(rcba_base + 0x3404);
139 if (!(val & 0x80)) {
140 err = 1;
141 } else {
142 val = val & 0x3;
143 force_hpet_address = 0xFED00000 | (val << 12);
146 if (err) {
147 force_hpet_address = 0;
148 iounmap(rcba_base);
149 dev_printk(KERN_DEBUG, &dev->dev,
150 "Failed to force enable HPET\n");
151 } else {
152 force_hpet_resume_type = ICH_FORCE_HPET_RESUME;
153 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
154 "0x%lx\n", force_hpet_address);
158 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB2_0,
159 ich_force_enable_hpet);
160 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_0,
161 ich_force_enable_hpet);
162 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_1,
163 ich_force_enable_hpet);
164 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_0,
165 ich_force_enable_hpet);
166 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_1,
167 ich_force_enable_hpet);
168 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_31,
169 ich_force_enable_hpet);
170 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_1,
171 ich_force_enable_hpet);
172 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_4,
173 ich_force_enable_hpet);
174 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH9_7,
175 ich_force_enable_hpet);
176 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, 0x3a16, /* ICH10 */
177 ich_force_enable_hpet);
179 static struct pci_dev *cached_dev;
181 static void hpet_print_force_info(void)
183 printk(KERN_INFO "HPET not enabled in BIOS. "
184 "You might try hpet=force boot option\n");
187 static void old_ich_force_hpet_resume(void)
189 u32 val;
190 u32 uninitialized_var(gen_cntl);
192 if (!force_hpet_address || !cached_dev)
193 return;
195 pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
196 gen_cntl &= (~(0x7 << 15));
197 gen_cntl |= (0x4 << 15);
199 pci_write_config_dword(cached_dev, 0xD0, gen_cntl);
200 pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
201 val = gen_cntl >> 15;
202 val &= 0x7;
203 if (val == 0x4)
204 printk(KERN_DEBUG "Force enabled HPET at resume\n");
205 else
206 BUG();
209 static void old_ich_force_enable_hpet(struct pci_dev *dev)
211 u32 val;
212 u32 uninitialized_var(gen_cntl);
214 if (hpet_address || force_hpet_address)
215 return;
217 pci_read_config_dword(dev, 0xD0, &gen_cntl);
219 * Bit 17 is HPET enable bit.
220 * Bit 16:15 control the HPET base address.
222 val = gen_cntl >> 15;
223 val &= 0x7;
224 if (val & 0x4) {
225 val &= 0x3;
226 force_hpet_address = 0xFED00000 | (val << 12);
227 dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
228 force_hpet_address);
229 return;
233 * HPET is disabled. Trying enabling at FED00000 and check
234 * whether it sticks
236 gen_cntl &= (~(0x7 << 15));
237 gen_cntl |= (0x4 << 15);
238 pci_write_config_dword(dev, 0xD0, gen_cntl);
240 pci_read_config_dword(dev, 0xD0, &gen_cntl);
242 val = gen_cntl >> 15;
243 val &= 0x7;
244 if (val & 0x4) {
245 /* HPET is enabled in HPTC. Just not reported by BIOS */
246 val &= 0x3;
247 force_hpet_address = 0xFED00000 | (val << 12);
248 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
249 "0x%lx\n", force_hpet_address);
250 cached_dev = dev;
251 force_hpet_resume_type = OLD_ICH_FORCE_HPET_RESUME;
252 return;
255 dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
259 * Undocumented chipset features. Make sure that the user enforced
260 * this.
262 static void old_ich_force_enable_hpet_user(struct pci_dev *dev)
264 if (hpet_force_user)
265 old_ich_force_enable_hpet(dev);
268 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB_1,
269 old_ich_force_enable_hpet_user);
270 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_0,
271 old_ich_force_enable_hpet_user);
272 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_12,
273 old_ich_force_enable_hpet_user);
274 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_0,
275 old_ich_force_enable_hpet_user);
276 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_12,
277 old_ich_force_enable_hpet_user);
278 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_0,
279 old_ich_force_enable_hpet);
280 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_12,
281 old_ich_force_enable_hpet);
284 static void vt8237_force_hpet_resume(void)
286 u32 val;
288 if (!force_hpet_address || !cached_dev)
289 return;
291 val = 0xfed00000 | 0x80;
292 pci_write_config_dword(cached_dev, 0x68, val);
294 pci_read_config_dword(cached_dev, 0x68, &val);
295 if (val & 0x80)
296 printk(KERN_DEBUG "Force enabled HPET at resume\n");
297 else
298 BUG();
301 static void vt8237_force_enable_hpet(struct pci_dev *dev)
303 u32 uninitialized_var(val);
305 if (hpet_address || force_hpet_address)
306 return;
308 if (!hpet_force_user) {
309 hpet_print_force_info();
310 return;
313 pci_read_config_dword(dev, 0x68, &val);
315 * Bit 7 is HPET enable bit.
316 * Bit 31:10 is HPET base address (contrary to what datasheet claims)
318 if (val & 0x80) {
319 force_hpet_address = (val & ~0x3ff);
320 dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
321 force_hpet_address);
322 return;
326 * HPET is disabled. Trying enabling at FED00000 and check
327 * whether it sticks
329 val = 0xfed00000 | 0x80;
330 pci_write_config_dword(dev, 0x68, val);
332 pci_read_config_dword(dev, 0x68, &val);
333 if (val & 0x80) {
334 force_hpet_address = (val & ~0x3ff);
335 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
336 "0x%lx\n", force_hpet_address);
337 cached_dev = dev;
338 force_hpet_resume_type = VT8237_FORCE_HPET_RESUME;
339 return;
342 dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
345 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235,
346 vt8237_force_enable_hpet);
347 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8237,
348 vt8237_force_enable_hpet);
349 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_CX700,
350 vt8237_force_enable_hpet);
352 static void ati_force_hpet_resume(void)
354 pci_write_config_dword(cached_dev, 0x14, 0xfed00000);
355 printk(KERN_DEBUG "Force enabled HPET at resume\n");
358 static u32 ati_ixp4x0_rev(struct pci_dev *dev)
360 int err = 0;
361 u32 d = 0;
362 u8 b = 0;
364 err = pci_read_config_byte(dev, 0xac, &b);
365 b &= ~(1<<5);
366 err |= pci_write_config_byte(dev, 0xac, b);
367 err |= pci_read_config_dword(dev, 0x70, &d);
368 d |= 1<<8;
369 err |= pci_write_config_dword(dev, 0x70, d);
370 err |= pci_read_config_dword(dev, 0x8, &d);
371 d &= 0xff;
372 dev_printk(KERN_DEBUG, &dev->dev, "SB4X0 revision 0x%x\n", d);
374 WARN_ON_ONCE(err);
376 return d;
379 static void ati_force_enable_hpet(struct pci_dev *dev)
381 u32 d, val;
382 u8 b;
384 if (hpet_address || force_hpet_address)
385 return;
387 if (!hpet_force_user) {
388 hpet_print_force_info();
389 return;
392 d = ati_ixp4x0_rev(dev);
393 if (d < 0x82)
394 return;
396 /* base address */
397 pci_write_config_dword(dev, 0x14, 0xfed00000);
398 pci_read_config_dword(dev, 0x14, &val);
400 /* enable interrupt */
401 outb(0x72, 0xcd6); b = inb(0xcd7);
402 b |= 0x1;
403 outb(0x72, 0xcd6); outb(b, 0xcd7);
404 outb(0x72, 0xcd6); b = inb(0xcd7);
405 if (!(b & 0x1))
406 return;
407 pci_read_config_dword(dev, 0x64, &d);
408 d |= (1<<10);
409 pci_write_config_dword(dev, 0x64, d);
410 pci_read_config_dword(dev, 0x64, &d);
411 if (!(d & (1<<10)))
412 return;
414 force_hpet_address = val;
415 force_hpet_resume_type = ATI_FORCE_HPET_RESUME;
416 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
417 force_hpet_address);
418 cached_dev = dev;
420 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_IXP400_SMBUS,
421 ati_force_enable_hpet);
424 * Undocumented chipset feature taken from LinuxBIOS.
426 static void nvidia_force_hpet_resume(void)
428 pci_write_config_dword(cached_dev, 0x44, 0xfed00001);
429 printk(KERN_DEBUG "Force enabled HPET at resume\n");
432 static void nvidia_force_enable_hpet(struct pci_dev *dev)
434 u32 uninitialized_var(val);
436 if (hpet_address || force_hpet_address)
437 return;
439 if (!hpet_force_user) {
440 hpet_print_force_info();
441 return;
444 pci_write_config_dword(dev, 0x44, 0xfed00001);
445 pci_read_config_dword(dev, 0x44, &val);
446 force_hpet_address = val & 0xfffffffe;
447 force_hpet_resume_type = NVIDIA_FORCE_HPET_RESUME;
448 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
449 force_hpet_address);
450 cached_dev = dev;
451 return;
454 /* ISA Bridges */
455 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0050,
456 nvidia_force_enable_hpet);
457 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0051,
458 nvidia_force_enable_hpet);
460 /* LPC bridges */
461 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0260,
462 nvidia_force_enable_hpet);
463 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0360,
464 nvidia_force_enable_hpet);
465 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0361,
466 nvidia_force_enable_hpet);
467 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0362,
468 nvidia_force_enable_hpet);
469 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0363,
470 nvidia_force_enable_hpet);
471 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0364,
472 nvidia_force_enable_hpet);
473 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0365,
474 nvidia_force_enable_hpet);
475 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0366,
476 nvidia_force_enable_hpet);
477 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0367,
478 nvidia_force_enable_hpet);
480 void force_hpet_resume(void)
482 switch (force_hpet_resume_type) {
483 case ICH_FORCE_HPET_RESUME:
484 ich_force_hpet_resume();
485 return;
486 case OLD_ICH_FORCE_HPET_RESUME:
487 old_ich_force_hpet_resume();
488 return;
489 case VT8237_FORCE_HPET_RESUME:
490 vt8237_force_hpet_resume();
491 return;
492 case NVIDIA_FORCE_HPET_RESUME:
493 nvidia_force_hpet_resume();
494 return;
495 case ATI_FORCE_HPET_RESUME:
496 ati_force_hpet_resume();
497 return;
498 default:
499 break;
504 * According to the datasheet e6xx systems have the HPET hardwired to
505 * 0xfed00000
507 static void e6xx_force_enable_hpet(struct pci_dev *dev)
509 if (hpet_address || force_hpet_address)
510 return;
512 force_hpet_address = 0xFED00000;
513 force_hpet_resume_type = NONE_FORCE_HPET_RESUME;
514 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
515 "0x%lx\n", force_hpet_address);
516 return;
518 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E6XX_CU,
519 e6xx_force_enable_hpet);
522 * HPET MSI on some boards (ATI SB700/SB800) has side effect on
523 * floppy DMA. Disable HPET MSI on such platforms.
524 * See erratum #27 (Misinterpreted MSI Requests May Result in
525 * Corrupted LPC DMA Data) in AMD Publication #46837,
526 * "SB700 Family Product Errata", Rev. 1.0, March 2010.
528 static void force_disable_hpet_msi(struct pci_dev *unused)
530 hpet_msi_disable = true;
533 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_SBX00_SMBUS,
534 force_disable_hpet_msi);
536 #endif
538 #if defined(CONFIG_PCI) && defined(CONFIG_NUMA)
539 /* Set correct numa_node information for AMD NB functions */
540 static void quirk_amd_nb_node(struct pci_dev *dev)
542 struct pci_dev *nb_ht;
543 unsigned int devfn;
544 u32 node;
545 u32 val;
547 devfn = PCI_DEVFN(PCI_SLOT(dev->devfn), 0);
548 nb_ht = pci_get_slot(dev->bus, devfn);
549 if (!nb_ht)
550 return;
552 pci_read_config_dword(nb_ht, 0x60, &val);
553 node = pcibus_to_node(dev->bus) | (val & 7);
555 * Some hardware may return an invalid node ID,
556 * so check it first:
558 if (node_online(node))
559 set_dev_node(&dev->dev, node);
560 pci_dev_put(nb_ht);
563 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB,
564 quirk_amd_nb_node);
565 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_ADDRMAP,
566 quirk_amd_nb_node);
567 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MEMCTL,
568 quirk_amd_nb_node);
569 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MISC,
570 quirk_amd_nb_node);
571 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_HT,
572 quirk_amd_nb_node);
573 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MAP,
574 quirk_amd_nb_node);
575 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_DRAM,
576 quirk_amd_nb_node);
577 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MISC,
578 quirk_amd_nb_node);
579 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_LINK,
580 quirk_amd_nb_node);
581 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F0,
582 quirk_amd_nb_node);
583 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F1,
584 quirk_amd_nb_node);
585 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F2,
586 quirk_amd_nb_node);
587 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F3,
588 quirk_amd_nb_node);
589 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F4,
590 quirk_amd_nb_node);
591 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F5,
592 quirk_amd_nb_node);
594 #endif
596 #ifdef CONFIG_PCI
598 * Processor does not ensure DRAM scrub read/write sequence
599 * is atomic wrt accesses to CC6 save state area. Therefore
600 * if a concurrent scrub read/write access is to same address
601 * the entry may appear as if it is not written. This quirk
602 * applies to Fam16h models 00h-0Fh
604 * See "Revision Guide" for AMD F16h models 00h-0fh,
605 * document 51810 rev. 3.04, Nov 2013
607 static void amd_disable_seq_and_redirect_scrub(struct pci_dev *dev)
609 u32 val;
612 * Suggested workaround:
613 * set D18F3x58[4:0] = 00h and set D18F3x5C[0] = 0b
615 pci_read_config_dword(dev, 0x58, &val);
616 if (val & 0x1F) {
617 val &= ~(0x1F);
618 pci_write_config_dword(dev, 0x58, val);
621 pci_read_config_dword(dev, 0x5C, &val);
622 if (val & BIT(0)) {
623 val &= ~BIT(0);
624 pci_write_config_dword(dev, 0x5c, val);
628 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_16H_NB_F3,
629 amd_disable_seq_and_redirect_scrub);
631 #if defined(CONFIG_X86_64) && defined(CONFIG_X86_MCE)
632 #include <linux/jump_label.h>
633 #include <asm/string_64.h>
635 /* Ivy Bridge, Haswell, Broadwell */
636 static void quirk_intel_brickland_xeon_ras_cap(struct pci_dev *pdev)
638 u32 capid0;
640 pci_read_config_dword(pdev, 0x84, &capid0);
642 if (capid0 & 0x10)
643 static_branch_inc(&mcsafe_key);
646 /* Skylake */
647 static void quirk_intel_purley_xeon_ras_cap(struct pci_dev *pdev)
649 u32 capid0, capid5;
651 pci_read_config_dword(pdev, 0x84, &capid0);
652 pci_read_config_dword(pdev, 0x98, &capid5);
655 * CAPID0{7:6} indicate whether this is an advanced RAS SKU
656 * CAPID5{8:5} indicate that various NVDIMM usage modes are
657 * enabled, so memory machine check recovery is also enabled.
659 if ((capid0 & 0xc0) == 0xc0 || (capid5 & 0x1e0))
660 static_branch_inc(&mcsafe_key);
663 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x0ec3, quirk_intel_brickland_xeon_ras_cap);
664 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2fc0, quirk_intel_brickland_xeon_ras_cap);
665 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x6fc0, quirk_intel_brickland_xeon_ras_cap);
666 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2083, quirk_intel_purley_xeon_ras_cap);
667 #endif
668 #endif
670 bool x86_apple_machine;
671 EXPORT_SYMBOL(x86_apple_machine);
673 void __init early_platform_quirks(void)
675 x86_apple_machine = dmi_match(DMI_SYS_VENDOR, "Apple Inc.") ||
676 dmi_match(DMI_SYS_VENDOR, "Apple Computer, Inc.");