2 * Intel e7xxx Memory Controller kernel module
3 * (C) 2003 Linux Networx (http://lnxi.com)
4 * This file may be distributed under the terms of the
5 * GNU General Public License.
7 * See "enum e7xxx_chips" below for supported chipsets
9 * Written by Thayne Harbaugh
10 * Based on work by Dan Hollis <goemon at anime dot net> and others.
11 * http://www.anime.net/~goemon/linux-ecc/
14 * http://www.intel.com/content/www/us/en/chipsets/e7501-chipset-memory-controller-hub-datasheet.html
17 * Eric Biederman (Linux Networx)
18 * Tom Zimmerman (Linux Networx)
19 * Jim Garlick (Lawrence Livermore National Labs)
20 * Dave Peterson (Lawrence Livermore National Labs)
21 * That One Guy (Some other place)
22 * Wang Zhenyu (intel.com)
24 * $Id: edac_e7xxx.c,v 1.5.2.9 2005/10/05 00:43:44 dsp_llnl Exp $
28 #include <linux/module.h>
29 #include <linux/init.h>
30 #include <linux/pci.h>
31 #include <linux/pci_ids.h>
32 #include <linux/edac.h>
33 #include "edac_core.h"
35 #define E7XXX_REVISION " Ver: 2.0.2"
36 #define EDAC_MOD_STR "e7xxx_edac"
38 #define e7xxx_printk(level, fmt, arg...) \
39 edac_printk(level, "e7xxx", fmt, ##arg)
41 #define e7xxx_mc_printk(mci, level, fmt, arg...) \
42 edac_mc_chipset_printk(mci, level, "e7xxx", fmt, ##arg)
44 #ifndef PCI_DEVICE_ID_INTEL_7205_0
45 #define PCI_DEVICE_ID_INTEL_7205_0 0x255d
46 #endif /* PCI_DEVICE_ID_INTEL_7205_0 */
48 #ifndef PCI_DEVICE_ID_INTEL_7205_1_ERR
49 #define PCI_DEVICE_ID_INTEL_7205_1_ERR 0x2551
50 #endif /* PCI_DEVICE_ID_INTEL_7205_1_ERR */
52 #ifndef PCI_DEVICE_ID_INTEL_7500_0
53 #define PCI_DEVICE_ID_INTEL_7500_0 0x2540
54 #endif /* PCI_DEVICE_ID_INTEL_7500_0 */
56 #ifndef PCI_DEVICE_ID_INTEL_7500_1_ERR
57 #define PCI_DEVICE_ID_INTEL_7500_1_ERR 0x2541
58 #endif /* PCI_DEVICE_ID_INTEL_7500_1_ERR */
60 #ifndef PCI_DEVICE_ID_INTEL_7501_0
61 #define PCI_DEVICE_ID_INTEL_7501_0 0x254c
62 #endif /* PCI_DEVICE_ID_INTEL_7501_0 */
64 #ifndef PCI_DEVICE_ID_INTEL_7501_1_ERR
65 #define PCI_DEVICE_ID_INTEL_7501_1_ERR 0x2541
66 #endif /* PCI_DEVICE_ID_INTEL_7501_1_ERR */
68 #ifndef PCI_DEVICE_ID_INTEL_7505_0
69 #define PCI_DEVICE_ID_INTEL_7505_0 0x2550
70 #endif /* PCI_DEVICE_ID_INTEL_7505_0 */
72 #ifndef PCI_DEVICE_ID_INTEL_7505_1_ERR
73 #define PCI_DEVICE_ID_INTEL_7505_1_ERR 0x2551
74 #endif /* PCI_DEVICE_ID_INTEL_7505_1_ERR */
76 #define E7XXX_NR_CSROWS 8 /* number of csrows */
77 #define E7XXX_NR_DIMMS 8 /* 2 channels, 4 dimms/channel */
79 /* E7XXX register addresses - device 0 function 0 */
80 #define E7XXX_DRB 0x60 /* DRAM row boundary register (8b) */
81 #define E7XXX_DRA 0x70 /* DRAM row attribute register (8b) */
83 * 31 Device width row 7 0=x8 1=x4
84 * 27 Device width row 6
85 * 23 Device width row 5
86 * 19 Device width row 4
87 * 15 Device width row 3
88 * 11 Device width row 2
89 * 7 Device width row 1
90 * 3 Device width row 0
92 #define E7XXX_DRC 0x7C /* DRAM controller mode reg (32b) */
94 * 22 Number channels 0=1,1=2
95 * 19:18 DRB Granularity 32/64MB
97 #define E7XXX_TOLM 0xC4 /* DRAM top of low memory reg (16b) */
98 #define E7XXX_REMAPBASE 0xC6 /* DRAM remap base address reg (16b) */
99 #define E7XXX_REMAPLIMIT 0xC8 /* DRAM remap limit address reg (16b) */
101 /* E7XXX register addresses - device 0 function 1 */
102 #define E7XXX_DRAM_FERR 0x80 /* DRAM first error register (8b) */
103 #define E7XXX_DRAM_NERR 0x82 /* DRAM next error register (8b) */
104 #define E7XXX_DRAM_CELOG_ADD 0xA0 /* DRAM first correctable memory */
105 /* error address register (32b) */
108 * 27:6 CE address (4k block 33:12)
111 #define E7XXX_DRAM_UELOG_ADD 0xB0 /* DRAM first uncorrectable memory */
112 /* error address register (32b) */
115 * 27:6 CE address (4k block 33:12)
118 #define E7XXX_DRAM_CELOG_SYNDROME 0xD0 /* DRAM first correctable memory */
119 /* error syndrome register (16b) */
129 struct pci_dev
*bridge_ck
;
133 const struct e7xxx_dev_info
*dev_info
;
136 struct e7xxx_dev_info
{
138 const char *ctl_name
;
141 struct e7xxx_error_info
{
145 u16 dram_celog_syndrome
;
149 static struct edac_pci_ctl_info
*e7xxx_pci
;
151 static const struct e7xxx_dev_info e7xxx_devs
[] = {
153 .err_dev
= PCI_DEVICE_ID_INTEL_7500_1_ERR
,
154 .ctl_name
= "E7500"},
156 .err_dev
= PCI_DEVICE_ID_INTEL_7501_1_ERR
,
157 .ctl_name
= "E7501"},
159 .err_dev
= PCI_DEVICE_ID_INTEL_7505_1_ERR
,
160 .ctl_name
= "E7505"},
162 .err_dev
= PCI_DEVICE_ID_INTEL_7205_1_ERR
,
163 .ctl_name
= "E7205"},
166 /* FIXME - is this valid for both SECDED and S4ECD4ED? */
167 static inline int e7xxx_find_channel(u16 syndrome
)
171 if ((syndrome
& 0xff00) == 0)
174 if ((syndrome
& 0x00ff) == 0)
177 if ((syndrome
& 0xf000) == 0 || (syndrome
& 0x0f00) == 0)
183 static unsigned long ctl_page_to_phys(struct mem_ctl_info
*mci
,
187 struct e7xxx_pvt
*pvt
= (struct e7xxx_pvt
*)mci
->pvt_info
;
191 if ((page
< pvt
->tolm
) ||
192 ((page
>= 0x100000) && (page
< pvt
->remapbase
)))
195 remap
= (page
- pvt
->tolm
) + pvt
->remapbase
;
197 if (remap
< pvt
->remaplimit
)
200 e7xxx_printk(KERN_ERR
, "Invalid page %lx - out of range\n", page
);
201 return pvt
->tolm
- 1;
204 static void process_ce(struct mem_ctl_info
*mci
, struct e7xxx_error_info
*info
)
212 /* read the error address */
213 error_1b
= info
->dram_celog_add
;
214 /* FIXME - should use PAGE_SHIFT */
215 page
= error_1b
>> 6; /* convert the address to 4k page */
216 /* read the syndrome */
217 syndrome
= info
->dram_celog_syndrome
;
218 /* FIXME - check for -1 */
219 row
= edac_mc_find_csrow_by_page(mci
, page
);
220 /* convert syndrome to channel */
221 channel
= e7xxx_find_channel(syndrome
);
222 edac_mc_handle_error(HW_EVENT_ERR_CORRECTED
, mci
, 1, page
, 0, syndrome
,
223 row
, channel
, -1, "e7xxx CE", "");
226 static void process_ce_no_info(struct mem_ctl_info
*mci
)
229 edac_mc_handle_error(HW_EVENT_ERR_CORRECTED
, mci
, 1, 0, 0, 0, -1, -1, -1,
230 "e7xxx CE log register overflow", "");
233 static void process_ue(struct mem_ctl_info
*mci
, struct e7xxx_error_info
*info
)
235 u32 error_2b
, block_page
;
239 /* read the error address */
240 error_2b
= info
->dram_uelog_add
;
241 /* FIXME - should use PAGE_SHIFT */
242 block_page
= error_2b
>> 6; /* convert to 4k address */
243 row
= edac_mc_find_csrow_by_page(mci
, block_page
);
245 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED
, mci
, 1, block_page
, 0, 0,
246 row
, -1, -1, "e7xxx UE", "");
249 static void process_ue_no_info(struct mem_ctl_info
*mci
)
253 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED
, mci
, 1, 0, 0, 0, -1, -1, -1,
254 "e7xxx UE log register overflow", "");
257 static void e7xxx_get_error_info(struct mem_ctl_info
*mci
,
258 struct e7xxx_error_info
*info
)
260 struct e7xxx_pvt
*pvt
;
262 pvt
= (struct e7xxx_pvt
*)mci
->pvt_info
;
263 pci_read_config_byte(pvt
->bridge_ck
, E7XXX_DRAM_FERR
, &info
->dram_ferr
);
264 pci_read_config_byte(pvt
->bridge_ck
, E7XXX_DRAM_NERR
, &info
->dram_nerr
);
266 if ((info
->dram_ferr
& 1) || (info
->dram_nerr
& 1)) {
267 pci_read_config_dword(pvt
->bridge_ck
, E7XXX_DRAM_CELOG_ADD
,
268 &info
->dram_celog_add
);
269 pci_read_config_word(pvt
->bridge_ck
,
270 E7XXX_DRAM_CELOG_SYNDROME
,
271 &info
->dram_celog_syndrome
);
274 if ((info
->dram_ferr
& 2) || (info
->dram_nerr
& 2))
275 pci_read_config_dword(pvt
->bridge_ck
, E7XXX_DRAM_UELOG_ADD
,
276 &info
->dram_uelog_add
);
278 if (info
->dram_ferr
& 3)
279 pci_write_bits8(pvt
->bridge_ck
, E7XXX_DRAM_FERR
, 0x03, 0x03);
281 if (info
->dram_nerr
& 3)
282 pci_write_bits8(pvt
->bridge_ck
, E7XXX_DRAM_NERR
, 0x03, 0x03);
285 static int e7xxx_process_error_info(struct mem_ctl_info
*mci
,
286 struct e7xxx_error_info
*info
,
293 /* decode and report errors */
294 if (info
->dram_ferr
& 1) { /* check first error correctable */
298 process_ce(mci
, info
);
301 if (info
->dram_ferr
& 2) { /* check first error uncorrectable */
305 process_ue(mci
, info
);
308 if (info
->dram_nerr
& 1) { /* check next error correctable */
312 if (info
->dram_ferr
& 1)
313 process_ce_no_info(mci
);
315 process_ce(mci
, info
);
319 if (info
->dram_nerr
& 2) { /* check next error uncorrectable */
323 if (info
->dram_ferr
& 2)
324 process_ue_no_info(mci
);
326 process_ue(mci
, info
);
333 static void e7xxx_check(struct mem_ctl_info
*mci
)
335 struct e7xxx_error_info info
;
338 e7xxx_get_error_info(mci
, &info
);
339 e7xxx_process_error_info(mci
, &info
, 1);
342 /* Return 1 if dual channel mode is active. Else return 0. */
343 static inline int dual_channel_active(u32 drc
, int dev_idx
)
345 return (dev_idx
== E7501
) ? ((drc
>> 22) & 0x1) : 1;
348 /* Return DRB granularity (0=32mb, 1=64mb). */
349 static inline int drb_granularity(u32 drc
, int dev_idx
)
351 /* only e7501 can be single channel */
352 return (dev_idx
== E7501
) ? ((drc
>> 18) & 0x3) : 1;
355 static void e7xxx_init_csrows(struct mem_ctl_info
*mci
, struct pci_dev
*pdev
,
356 int dev_idx
, u32 drc
)
358 unsigned long last_cumul_size
;
361 u32 dra
, cumul_size
, nr_pages
;
362 int drc_chan
, drc_drbg
, drc_ddim
, mem_dev
;
363 struct csrow_info
*csrow
;
364 struct dimm_info
*dimm
;
365 enum edac_type edac_mode
;
367 pci_read_config_dword(pdev
, E7XXX_DRA
, &dra
);
368 drc_chan
= dual_channel_active(drc
, dev_idx
);
369 drc_drbg
= drb_granularity(drc
, dev_idx
);
370 drc_ddim
= (drc
>> 20) & 0x3;
373 /* The dram row boundary (DRB) reg values are boundary address
374 * for each DRAM row with a granularity of 32 or 64MB (single/dual
375 * channel operation). DRB regs are cumulative; therefore DRB7 will
376 * contain the total memory contained in all eight rows.
378 for (index
= 0; index
< mci
->nr_csrows
; index
++) {
379 /* mem_dev 0=x8, 1=x4 */
380 mem_dev
= (dra
>> (index
* 4 + 3)) & 0x1;
381 csrow
= mci
->csrows
[index
];
383 pci_read_config_byte(pdev
, E7XXX_DRB
+ index
, &value
);
384 /* convert a 64 or 32 MiB DRB to a page size. */
385 cumul_size
= value
<< (25 + drc_drbg
- PAGE_SHIFT
);
386 edac_dbg(3, "(%d) cumul_size 0x%x\n", index
, cumul_size
);
387 if (cumul_size
== last_cumul_size
)
388 continue; /* not populated */
390 csrow
->first_page
= last_cumul_size
;
391 csrow
->last_page
= cumul_size
- 1;
392 nr_pages
= cumul_size
- last_cumul_size
;
393 last_cumul_size
= cumul_size
;
396 * if single channel or x8 devices then SECDED
397 * if dual channel and x4 then S4ECD4ED
400 if (drc_chan
&& mem_dev
) {
401 edac_mode
= EDAC_S4ECD4ED
;
402 mci
->edac_cap
|= EDAC_FLAG_S4ECD4ED
;
404 edac_mode
= EDAC_SECDED
;
405 mci
->edac_cap
|= EDAC_FLAG_SECDED
;
408 edac_mode
= EDAC_NONE
;
410 for (j
= 0; j
< drc_chan
+ 1; j
++) {
411 dimm
= csrow
->channels
[j
]->dimm
;
413 dimm
->nr_pages
= nr_pages
/ (drc_chan
+ 1);
414 dimm
->grain
= 1 << 12; /* 4KiB - resolution of CELOG */
415 dimm
->mtype
= MEM_RDDR
; /* only one type supported */
416 dimm
->dtype
= mem_dev
? DEV_X4
: DEV_X8
;
417 dimm
->edac_mode
= edac_mode
;
422 static int e7xxx_probe1(struct pci_dev
*pdev
, int dev_idx
)
425 struct mem_ctl_info
*mci
= NULL
;
426 struct edac_mc_layer layers
[2];
427 struct e7xxx_pvt
*pvt
= NULL
;
430 struct e7xxx_error_info discard
;
432 edac_dbg(0, "mci\n");
434 pci_read_config_dword(pdev
, E7XXX_DRC
, &drc
);
436 drc_chan
= dual_channel_active(drc
, dev_idx
);
438 * According with the datasheet, this device has a maximum of
439 * 4 DIMMS per channel, either single-rank or dual-rank. So, the
440 * total amount of dimms is 8 (E7XXX_NR_DIMMS).
441 * That means that the DIMM is mapped as CSROWs, and the channel
442 * will map the rank. So, an error to either channel should be
443 * attributed to the same dimm.
445 layers
[0].type
= EDAC_MC_LAYER_CHIP_SELECT
;
446 layers
[0].size
= E7XXX_NR_CSROWS
;
447 layers
[0].is_virt_csrow
= true;
448 layers
[1].type
= EDAC_MC_LAYER_CHANNEL
;
449 layers
[1].size
= drc_chan
+ 1;
450 layers
[1].is_virt_csrow
= false;
451 mci
= edac_mc_alloc(0, ARRAY_SIZE(layers
), layers
, sizeof(*pvt
));
455 edac_dbg(3, "init mci\n");
456 mci
->mtype_cap
= MEM_FLAG_RDDR
;
457 mci
->edac_ctl_cap
= EDAC_FLAG_NONE
| EDAC_FLAG_SECDED
|
459 /* FIXME - what if different memory types are in different csrows? */
460 mci
->mod_name
= EDAC_MOD_STR
;
461 mci
->mod_ver
= E7XXX_REVISION
;
462 mci
->pdev
= &pdev
->dev
;
463 edac_dbg(3, "init pvt\n");
464 pvt
= (struct e7xxx_pvt
*)mci
->pvt_info
;
465 pvt
->dev_info
= &e7xxx_devs
[dev_idx
];
466 pvt
->bridge_ck
= pci_get_device(PCI_VENDOR_ID_INTEL
,
467 pvt
->dev_info
->err_dev
, pvt
->bridge_ck
);
469 if (!pvt
->bridge_ck
) {
470 e7xxx_printk(KERN_ERR
, "error reporting device not found:"
471 "vendor %x device 0x%x (broken BIOS?)\n",
472 PCI_VENDOR_ID_INTEL
, e7xxx_devs
[dev_idx
].err_dev
);
476 edac_dbg(3, "more mci init\n");
477 mci
->ctl_name
= pvt
->dev_info
->ctl_name
;
478 mci
->dev_name
= pci_name(pdev
);
479 mci
->edac_check
= e7xxx_check
;
480 mci
->ctl_page_to_phys
= ctl_page_to_phys
;
481 e7xxx_init_csrows(mci
, pdev
, dev_idx
, drc
);
482 mci
->edac_cap
|= EDAC_FLAG_NONE
;
483 edac_dbg(3, "tolm, remapbase, remaplimit\n");
484 /* load the top of low memory, remap base, and remap limit vars */
485 pci_read_config_word(pdev
, E7XXX_TOLM
, &pci_data
);
486 pvt
->tolm
= ((u32
) pci_data
) << 4;
487 pci_read_config_word(pdev
, E7XXX_REMAPBASE
, &pci_data
);
488 pvt
->remapbase
= ((u32
) pci_data
) << 14;
489 pci_read_config_word(pdev
, E7XXX_REMAPLIMIT
, &pci_data
);
490 pvt
->remaplimit
= ((u32
) pci_data
) << 14;
491 e7xxx_printk(KERN_INFO
,
492 "tolm = %x, remapbase = %x, remaplimit = %x\n", pvt
->tolm
,
493 pvt
->remapbase
, pvt
->remaplimit
);
495 /* clear any pending errors, or initial state bits */
496 e7xxx_get_error_info(mci
, &discard
);
498 /* Here we assume that we will never see multiple instances of this
499 * type of memory controller. The ID is therefore hardcoded to 0.
501 if (edac_mc_add_mc(mci
)) {
502 edac_dbg(3, "failed edac_mc_add_mc()\n");
506 /* allocating generic PCI control info */
507 e7xxx_pci
= edac_pci_create_generic_ctl(&pdev
->dev
, EDAC_MOD_STR
);
510 "%s(): Unable to create PCI control\n",
513 "%s(): PCI error report via EDAC not setup\n",
517 /* get this far and it's successful */
518 edac_dbg(3, "success\n");
522 pci_dev_put(pvt
->bridge_ck
);
530 /* returns count (>= 0), or negative on error */
531 static int e7xxx_init_one(struct pci_dev
*pdev
, const struct pci_device_id
*ent
)
535 /* wake up and enable device */
536 return pci_enable_device(pdev
) ?
537 -EIO
: e7xxx_probe1(pdev
, ent
->driver_data
);
540 static void e7xxx_remove_one(struct pci_dev
*pdev
)
542 struct mem_ctl_info
*mci
;
543 struct e7xxx_pvt
*pvt
;
548 edac_pci_release_generic_ctl(e7xxx_pci
);
550 if ((mci
= edac_mc_del_mc(&pdev
->dev
)) == NULL
)
553 pvt
= (struct e7xxx_pvt
*)mci
->pvt_info
;
554 pci_dev_put(pvt
->bridge_ck
);
558 static const struct pci_device_id e7xxx_pci_tbl
[] = {
560 PCI_VEND_DEV(INTEL
, 7205_0
), PCI_ANY_ID
, PCI_ANY_ID
, 0, 0,
563 PCI_VEND_DEV(INTEL
, 7500_0
), PCI_ANY_ID
, PCI_ANY_ID
, 0, 0,
566 PCI_VEND_DEV(INTEL
, 7501_0
), PCI_ANY_ID
, PCI_ANY_ID
, 0, 0,
569 PCI_VEND_DEV(INTEL
, 7505_0
), PCI_ANY_ID
, PCI_ANY_ID
, 0, 0,
573 } /* 0 terminated list. */
576 MODULE_DEVICE_TABLE(pci
, e7xxx_pci_tbl
);
578 static struct pci_driver e7xxx_driver
= {
579 .name
= EDAC_MOD_STR
,
580 .probe
= e7xxx_init_one
,
581 .remove
= e7xxx_remove_one
,
582 .id_table
= e7xxx_pci_tbl
,
585 static int __init
e7xxx_init(void)
587 /* Ensure that the OPSTATE is set correctly for POLL or NMI */
590 return pci_register_driver(&e7xxx_driver
);
593 static void __exit
e7xxx_exit(void)
595 pci_unregister_driver(&e7xxx_driver
);
598 module_init(e7xxx_init
);
599 module_exit(e7xxx_exit
);
601 MODULE_LICENSE("GPL");
602 MODULE_AUTHOR("Linux Networx (http://lnxi.com) Thayne Harbaugh et al\n"
603 "Based on.work by Dan Hollis et al");
604 MODULE_DESCRIPTION("MC support for Intel e7xxx memory controllers");
605 module_param(edac_op_state
, int, 0444);
606 MODULE_PARM_DESC(edac_op_state
, "EDAC Error Reporting state: 0=Poll,1=NMI");