1 /*
2 * Intel e752x Memory Controller kernel module
3 * (C) 2004 Linux Networx (http://lnxi.com)
4 * This file may be distributed under the terms of the
5 * GNU General Public License.
6 *
7 * See "enum e752x_chips" below for supported chipsets
8 *
9 * Written by Tom Zimmerman
10 *
11 * Contributors:
12 * Thayne Harbaugh at realmsys.com (?)
13 * Wang Zhenyu at intel.com
14 * Dave Jiang at mvista.com
15 *
16 * $Id: edac_e752x.c,v 1.5.2.11 2005/10/05 00:43:44 dsp_llnl Exp $
17 *
18 */
19
20 #include <linux/module.h>
21 #include <linux/init.h>
22 #include <linux/pci.h>
23 #include <linux/pci_ids.h>
24 #include <linux/edac.h>
25 #include "edac_core.h"
26
27 #define E752X_REVISION " Ver: 2.0.2 " __DATE__
28 #define EDAC_MOD_STR "e752x_edac"
29
30 static int report_non_memory_errors;
31 static int force_function_unhide;
32 static int sysbus_parity = -1;
33
34 static struct edac_pci_ctl_info *e752x_pci;
35
36 #define e752x_printk(level, fmt, arg...) \
37 edac_printk(level, "e752x", fmt, ##arg)
38
39 #define e752x_mc_printk(mci, level, fmt, arg...) \
40 edac_mc_chipset_printk(mci, level, "e752x", fmt, ##arg)
41
42 #ifndef PCI_DEVICE_ID_INTEL_7520_0
43 #define PCI_DEVICE_ID_INTEL_7520_0 0x3590
44 #endif /* PCI_DEVICE_ID_INTEL_7520_0 */
45
46 #ifndef PCI_DEVICE_ID_INTEL_7520_1_ERR
47 #define PCI_DEVICE_ID_INTEL_7520_1_ERR 0x3591
48 #endif /* PCI_DEVICE_ID_INTEL_7520_1_ERR */
49
50 #ifndef PCI_DEVICE_ID_INTEL_7525_0
51 #define PCI_DEVICE_ID_INTEL_7525_0 0x359E
52 #endif /* PCI_DEVICE_ID_INTEL_7525_0 */
53
54 #ifndef PCI_DEVICE_ID_INTEL_7525_1_ERR
55 #define PCI_DEVICE_ID_INTEL_7525_1_ERR 0x3593
56 #endif /* PCI_DEVICE_ID_INTEL_7525_1_ERR */
57
58 #ifndef PCI_DEVICE_ID_INTEL_7320_0
59 #define PCI_DEVICE_ID_INTEL_7320_0 0x3592
60 #endif /* PCI_DEVICE_ID_INTEL_7320_0 */
61
62 #ifndef PCI_DEVICE_ID_INTEL_7320_1_ERR
63 #define PCI_DEVICE_ID_INTEL_7320_1_ERR 0x3593
64 #endif /* PCI_DEVICE_ID_INTEL_7320_1_ERR */
65
66 #ifndef PCI_DEVICE_ID_INTEL_3100_0
67 #define PCI_DEVICE_ID_INTEL_3100_0 0x35B0
68 #endif /* PCI_DEVICE_ID_INTEL_3100_0 */
69
70 #ifndef PCI_DEVICE_ID_INTEL_3100_1_ERR
71 #define PCI_DEVICE_ID_INTEL_3100_1_ERR 0x35B1
72 #endif /* PCI_DEVICE_ID_INTEL_3100_1_ERR */
73
74 #define E752X_NR_CSROWS 8 /* number of csrows */
75
76 /* E752X register addresses - device 0 function 0 */
77 #define E752X_MCHSCRB 0x52 /* Memory Scrub register (16b) */
78 /*
79 * 6:5 Scrub Completion Count
80 * 3:2 Scrub Rate (i3100 only)
81 * 01=fast 10=normal
82 * 1:0 Scrub Mode enable
83 * 00=off 10=on
84 */
85 #define E752X_DRB 0x60 /* DRAM row boundary register (8b) */
86 #define E752X_DRA 0x70 /* DRAM row attribute register (8b) */
87 /*
88 * 31:30 Device width row 7
89 * 01=x8 10=x4 11=x8 DDR2
90 * 27:26 Device width row 6
91 * 23:22 Device width row 5
92 * 19:20 Device width row 4
93 * 15:14 Device width row 3
94 * 11:10 Device width row 2
95 * 7:6 Device width row 1
96 * 3:2 Device width row 0
97 */
98 #define E752X_DRC 0x7C /* DRAM controller mode reg (32b) */
99 /* FIXME:IS THIS RIGHT? */
100 /*
101 * 22 Number channels 0=1,1=2
102 * 19:18 DRB Granularity 32/64MB
103 */
104 #define E752X_DRM 0x80 /* Dimm mapping register */
105 #define E752X_DDRCSR 0x9A /* DDR control and status reg (16b) */
106 /*
107 * 14:12 1 single A, 2 single B, 3 dual
108 */
109 #define E752X_TOLM 0xC4 /* DRAM top of low memory reg (16b) */
110 #define E752X_REMAPBASE 0xC6 /* DRAM remap base address reg (16b) */
111 #define E752X_REMAPLIMIT 0xC8 /* DRAM remap limit address reg (16b) */
112 #define E752X_REMAPOFFSET 0xCA /* DRAM remap limit offset reg (16b) */
113
114 /* E752X register addresses - device 0 function 1 */
115 #define E752X_FERR_GLOBAL 0x40 /* Global first error register (32b) */
116 #define E752X_NERR_GLOBAL 0x44 /* Global next error register (32b) */
117 #define E752X_HI_FERR 0x50 /* Hub interface first error reg (8b) */
118 #define E752X_HI_NERR 0x52 /* Hub interface next error reg (8b) */
119 #define E752X_HI_ERRMASK 0x54 /* Hub interface error mask reg (8b) */
120 #define E752X_HI_SMICMD 0x5A /* Hub interface SMI command reg (8b) */
121 #define E752X_SYSBUS_FERR 0x60 /* System buss first error reg (16b) */
122 #define E752X_SYSBUS_NERR 0x62 /* System buss next error reg (16b) */
123 #define E752X_SYSBUS_ERRMASK 0x64 /* System buss error mask reg (16b) */
124 #define E752X_SYSBUS_SMICMD 0x6A /* System buss SMI command reg (16b) */
125 #define E752X_BUF_FERR 0x70 /* Memory buffer first error reg (8b) */
126 #define E752X_BUF_NERR 0x72 /* Memory buffer next error reg (8b) */
127 #define E752X_BUF_ERRMASK 0x74 /* Memory buffer error mask reg (8b) */
128 #define E752X_BUF_SMICMD 0x7A /* Memory buffer SMI cmd reg (8b) */
129 #define E752X_DRAM_FERR 0x80 /* DRAM first error register (16b) */
130 #define E752X_DRAM_NERR 0x82 /* DRAM next error register (16b) */
131 #define E752X_DRAM_ERRMASK 0x84 /* DRAM error mask register (8b) */
132 #define E752X_DRAM_SMICMD 0x8A /* DRAM SMI command register (8b) */
133 #define E752X_DRAM_RETR_ADD 0xAC /* DRAM Retry address register (32b) */
134 #define E752X_DRAM_SEC1_ADD 0xA0 /* DRAM first correctable memory */
135 /* error address register (32b) */
136 /*
137 * 31 Reserved
138 * 30:2 CE address (64 byte block 34:6
139 * 1 Reserved
140 * 0 HiLoCS
141 */
142 #define E752X_DRAM_SEC2_ADD 0xC8 /* DRAM first correctable memory */
143 /* error address register (32b) */
144 /*
145 * 31 Reserved
146 * 30:2 CE address (64 byte block 34:6)
147 * 1 Reserved
148 * 0 HiLoCS
149 */
150 #define E752X_DRAM_DED_ADD 0xA4 /* DRAM first uncorrectable memory */
151 /* error address register (32b) */
152 /*
153 * 31 Reserved
154 * 30:2 CE address (64 byte block 34:6)
155 * 1 Reserved
156 * 0 HiLoCS
157 */
158 #define E752X_DRAM_SCRB_ADD 0xA8 /* DRAM 1st uncorrectable scrub mem */
159 /* error address register (32b) */
160 /*
161 * 31 Reserved
162 * 30:2 CE address (64 byte block 34:6
163 * 1 Reserved
164 * 0 HiLoCS
165 */
166 #define E752X_DRAM_SEC1_SYNDROME 0xC4 /* DRAM first correctable memory */
167 /* error syndrome register (16b) */
168 #define E752X_DRAM_SEC2_SYNDROME 0xC6 /* DRAM second correctable memory */
169 /* error syndrome register (16b) */
170 #define E752X_DEVPRES1 0xF4 /* Device Present 1 register (8b) */
171
172 /* 3100 IMCH specific register addresses - device 0 function 1 */
173 #define I3100_NSI_FERR 0x48 /* NSI first error reg (32b) */
174 #define I3100_NSI_NERR 0x4C /* NSI next error reg (32b) */
175 #define I3100_NSI_SMICMD 0x54 /* NSI SMI command register (32b) */
176 #define I3100_NSI_EMASK 0x90 /* NSI error mask register (32b) */
177
178 /* ICH5R register addresses - device 30 function 0 */
179 #define ICH5R_PCI_STAT 0x06 /* PCI status register (16b) */
180 #define ICH5R_PCI_2ND_STAT 0x1E /* PCI status secondary reg (16b) */
181 #define ICH5R_PCI_BRIDGE_CTL 0x3E /* PCI bridge control register (16b) */
182
183 enum e752x_chips {
184 E7520 = 0,
185 E7525 = 1,
186 E7320 = 2,
187 I3100 = 3
188 };
189
190 struct e752x_pvt {
191 struct pci_dev *bridge_ck;
192 struct pci_dev *dev_d0f0;
193 struct pci_dev *dev_d0f1;
194 u32 tolm;
195 u32 remapbase;
196 u32 remaplimit;
197 int mc_symmetric;
198 u8 map[8];
199 int map_type;
200 const struct e752x_dev_info *dev_info;
201 };
202
203 struct e752x_dev_info {
204 u16 err_dev;
205 u16 ctl_dev;
206 const char *ctl_name;
207 };
208
209 struct e752x_error_info {
210 u32 ferr_global;
211 u32 nerr_global;
212 u32 nsi_ferr; /* 3100 only */
213 u32 nsi_nerr; /* 3100 only */
214 u8 hi_ferr; /* all but 3100 */
215 u8 hi_nerr; /* all but 3100 */
216 u16 sysbus_ferr;
217 u16 sysbus_nerr;
218 u8 buf_ferr;
219 u8 buf_nerr;
220 u16 dram_ferr;
221 u16 dram_nerr;
222 u32 dram_sec1_add;
223 u32 dram_sec2_add;
224 u16 dram_sec1_syndrome;
225 u16 dram_sec2_syndrome;
226 u32 dram_ded_add;
227 u32 dram_scrb_add;
228 u32 dram_retr_add;
229 };
230
231 static const struct e752x_dev_info e752x_devs[] = {
232 [E7520] = {
233 .err_dev = PCI_DEVICE_ID_INTEL_7520_1_ERR,
234 .ctl_dev = PCI_DEVICE_ID_INTEL_7520_0,
235 .ctl_name = "E7520"},
236 [E7525] = {
237 .err_dev = PCI_DEVICE_ID_INTEL_7525_1_ERR,
238 .ctl_dev = PCI_DEVICE_ID_INTEL_7525_0,
239 .ctl_name = "E7525"},
240 [E7320] = {
241 .err_dev = PCI_DEVICE_ID_INTEL_7320_1_ERR,
242 .ctl_dev = PCI_DEVICE_ID_INTEL_7320_0,
243 .ctl_name = "E7320"},
244 [I3100] = {
245 .err_dev = PCI_DEVICE_ID_INTEL_3100_1_ERR,
246 .ctl_dev = PCI_DEVICE_ID_INTEL_3100_0,
247 .ctl_name = "3100"},
248 };
249
250 /* Valid scrub rates for the e752x/3100 hardware memory scrubber. We
251 * map the scrubbing bandwidth to a hardware register value. The 'set'
252 * operation finds the 'matching or higher value'. Note that scrubbing
253 * on the e752x can only be enabled/disabled. The 3100 supports
254 * a normal and fast mode.
255 */
256
257 #define SDRATE_EOT 0xFFFFFFFF
258
259 struct scrubrate {
260 u32 bandwidth; /* bandwidth consumed by scrubbing in bytes/sec */
261 u16 scrubval; /* register value for scrub rate */
262 };
263
264 /* Rate below assumes same performance as i3100 using PC3200 DDR2 in
265 * normal mode. e752x bridges don't support choosing normal or fast mode,
266 * so the scrubbing bandwidth value isn't all that important - scrubbing is
267 * either on or off.
268 */
269 static const struct scrubrate scrubrates_e752x[] = {
270 {0, 0x00}, /* Scrubbing Off */
271 {500000, 0x02}, /* Scrubbing On */
272 {SDRATE_EOT, 0x00} /* End of Table */
273 };
274
275 /* Fast mode: 2 GByte PC3200 DDR2 scrubbed in 33s = 63161283 bytes/s
276 * Normal mode: 125 (32000 / 256) times slower than fast mode.
277 */
278 static const struct scrubrate scrubrates_i3100[] = {
279 {0, 0x00}, /* Scrubbing Off */
280 {500000, 0x0a}, /* Normal mode - 32k clocks */
281 {62500000, 0x06}, /* Fast mode - 256 clocks */
282 {SDRATE_EOT, 0x00} /* End of Table */
283 };
284
ctl_page_to_phys(struct mem_ctl_info * mci,unsigned long page)285 static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci,
286 unsigned long page)
287 {
288 u32 remap;
289 struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
290
291 debugf3("%s()\n", __func__);
292
293 if (page < pvt->tolm)
294 return page;
295
296 if ((page >= 0x100000) && (page < pvt->remapbase))
297 return page;
298
299 remap = (page - pvt->tolm) + pvt->remapbase;
300
301 if (remap < pvt->remaplimit)
302 return remap;
303
304 e752x_printk(KERN_ERR, "Invalid page %lx - out of range\n", page);
305 return pvt->tolm - 1;
306 }
307
do_process_ce(struct mem_ctl_info * mci,u16 error_one,u32 sec1_add,u16 sec1_syndrome)308 static void do_process_ce(struct mem_ctl_info *mci, u16 error_one,
309 u32 sec1_add, u16 sec1_syndrome)
310 {
311 u32 page;
312 int row;
313 int channel;
314 int i;
315 struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
316
317 debugf3("%s()\n", __func__);
318
319 /* convert the addr to 4k page */
320 page = sec1_add >> (PAGE_SHIFT - 4);
321
322 /* FIXME - check for -1 */
323 if (pvt->mc_symmetric) {
324 /* chip select are bits 14 & 13 */
325 row = ((page >> 1) & 3);
326 e752x_printk(KERN_WARNING,
327 "Test row %d Table %d %d %d %d %d %d %d %d\n", row,
328 pvt->map[0], pvt->map[1], pvt->map[2], pvt->map[3],
329 pvt->map[4], pvt->map[5], pvt->map[6],
330 pvt->map[7]);
331
332 /* test for channel remapping */
333 for (i = 0; i < 8; i++) {
334 if (pvt->map[i] == row)
335 break;
336 }
337
338 e752x_printk(KERN_WARNING, "Test computed row %d\n", i);
339
340 if (i < 8)
341 row = i;
342 else
343 e752x_mc_printk(mci, KERN_WARNING,
344 "row %d not found in remap table\n",
345 row);
346 } else
347 row = edac_mc_find_csrow_by_page(mci, page);
348
349 /* 0 = channel A, 1 = channel B */
350 channel = !(error_one & 1);
351
352 /* e752x mc reads 34:6 of the DRAM linear address */
353 edac_mc_handle_ce(mci, page, offset_in_page(sec1_add << 4),
354 sec1_syndrome, row, channel, "e752x CE");
355 }
356
process_ce(struct mem_ctl_info * mci,u16 error_one,u32 sec1_add,u16 sec1_syndrome,int * error_found,int handle_error)357 static inline void process_ce(struct mem_ctl_info *mci, u16 error_one,
358 u32 sec1_add, u16 sec1_syndrome, int *error_found,
359 int handle_error)
360 {
361 *error_found = 1;
362
363 if (handle_error)
364 do_process_ce(mci, error_one, sec1_add, sec1_syndrome);
365 }
366
do_process_ue(struct mem_ctl_info * mci,u16 error_one,u32 ded_add,u32 scrb_add)367 static void do_process_ue(struct mem_ctl_info *mci, u16 error_one,
368 u32 ded_add, u32 scrb_add)
369 {
370 u32 error_2b, block_page;
371 int row;
372 struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
373
374 debugf3("%s()\n", __func__);
375
376 if (error_one & 0x0202) {
377 error_2b = ded_add;
378
379 /* convert to 4k address */
380 block_page = error_2b >> (PAGE_SHIFT - 4);
381
382 row = pvt->mc_symmetric ?
383 /* chip select are bits 14 & 13 */
384 ((block_page >> 1) & 3) :
385 edac_mc_find_csrow_by_page(mci, block_page);
386
387 /* e752x mc reads 34:6 of the DRAM linear address */
388 edac_mc_handle_ue(mci, block_page,
389 offset_in_page(error_2b << 4),
390 row, "e752x UE from Read");
391 }
392 if (error_one & 0x0404) {
393 error_2b = scrb_add;
394
395 /* convert to 4k address */
396 block_page = error_2b >> (PAGE_SHIFT - 4);
397
398 row = pvt->mc_symmetric ?
399 /* chip select are bits 14 & 13 */
400 ((block_page >> 1) & 3) :
401 edac_mc_find_csrow_by_page(mci, block_page);
402
403 /* e752x mc reads 34:6 of the DRAM linear address */
404 edac_mc_handle_ue(mci, block_page,
405 offset_in_page(error_2b << 4),
406 row, "e752x UE from Scruber");
407 }
408 }
409
process_ue(struct mem_ctl_info * mci,u16 error_one,u32 ded_add,u32 scrb_add,int * error_found,int handle_error)410 static inline void process_ue(struct mem_ctl_info *mci, u16 error_one,
411 u32 ded_add, u32 scrb_add, int *error_found,
412 int handle_error)
413 {
414 *error_found = 1;
415
416 if (handle_error)
417 do_process_ue(mci, error_one, ded_add, scrb_add);
418 }
419
process_ue_no_info_wr(struct mem_ctl_info * mci,int * error_found,int handle_error)420 static inline void process_ue_no_info_wr(struct mem_ctl_info *mci,
421 int *error_found, int handle_error)
422 {
423 *error_found = 1;
424
425 if (!handle_error)
426 return;
427
428 debugf3("%s()\n", __func__);
429 edac_mc_handle_ue_no_info(mci, "e752x UE log memory write");
430 }
431
do_process_ded_retry(struct mem_ctl_info * mci,u16 error,u32 retry_add)432 static void do_process_ded_retry(struct mem_ctl_info *mci, u16 error,
433 u32 retry_add)
434 {
435 u32 error_1b, page;
436 int row;
437 struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
438
439 error_1b = retry_add;
440 page = error_1b >> (PAGE_SHIFT - 4); /* convert the addr to 4k page */
441
442 /* chip select are bits 14 & 13 */
443 row = pvt->mc_symmetric ? ((page >> 1) & 3) :
444 edac_mc_find_csrow_by_page(mci, page);
445
446 e752x_mc_printk(mci, KERN_WARNING,
447 "CE page 0x%lx, row %d : Memory read retry\n",
448 (long unsigned int)page, row);
449 }
450
process_ded_retry(struct mem_ctl_info * mci,u16 error,u32 retry_add,int * error_found,int handle_error)451 static inline void process_ded_retry(struct mem_ctl_info *mci, u16 error,
452 u32 retry_add, int *error_found,
453 int handle_error)
454 {
455 *error_found = 1;
456
457 if (handle_error)
458 do_process_ded_retry(mci, error, retry_add);
459 }
460
process_threshold_ce(struct mem_ctl_info * mci,u16 error,int * error_found,int handle_error)461 static inline void process_threshold_ce(struct mem_ctl_info *mci, u16 error,
462 int *error_found, int handle_error)
463 {
464 *error_found = 1;
465
466 if (handle_error)
467 e752x_mc_printk(mci, KERN_WARNING, "Memory threshold CE\n");
468 }
469
470 static char *global_message[11] = {
471 "PCI Express C1",
472 "PCI Express C",
473 "PCI Express B1",
474 "PCI Express B",
475 "PCI Express A1",
476 "PCI Express A",
477 "DMA Controller",
478 "HUB or NS Interface",
479 "System Bus",
480 "DRAM Controller", /* 9th entry */
481 "Internal Buffer"
482 };
483
484 #define DRAM_ENTRY 9
485
486 static char *fatal_message[2] = { "Non-Fatal ", "Fatal " };
487
do_global_error(int fatal,u32 errors)488 static void do_global_error(int fatal, u32 errors)
489 {
490 int i;
491
492 for (i = 0; i < 11; i++) {
493 if (errors & (1 << i)) {
494 /* If the error is from DRAM Controller OR
495 * we are to report ALL errors, then
496 * report the error
497 */
498 if ((i == DRAM_ENTRY) || report_non_memory_errors)
499 e752x_printk(KERN_WARNING, "%sError %s\n",
500 fatal_message[fatal],
501 global_message[i]);
502 }
503 }
504 }
505
global_error(int fatal,u32 errors,int * error_found,int handle_error)506 static inline void global_error(int fatal, u32 errors, int *error_found,
507 int handle_error)
508 {
509 *error_found = 1;
510
511 if (handle_error)
512 do_global_error(fatal, errors);
513 }
514
515 static char *hub_message[7] = {
516 "HI Address or Command Parity", "HI Illegal Access",
517 "HI Internal Parity", "Out of Range Access",
518 "HI Data Parity", "Enhanced Config Access",
519 "Hub Interface Target Abort"
520 };
521
do_hub_error(int fatal,u8 errors)522 static void do_hub_error(int fatal, u8 errors)
523 {
524 int i;
525
526 for (i = 0; i < 7; i++) {
527 if (errors & (1 << i))
528 e752x_printk(KERN_WARNING, "%sError %s\n",
529 fatal_message[fatal], hub_message[i]);
530 }
531 }
532
hub_error(int fatal,u8 errors,int * error_found,int handle_error)533 static inline void hub_error(int fatal, u8 errors, int *error_found,
534 int handle_error)
535 {
536 *error_found = 1;
537
538 if (handle_error)
539 do_hub_error(fatal, errors);
540 }
541
542 #define NSI_FATAL_MASK 0x0c080081
543 #define NSI_NON_FATAL_MASK 0x23a0ba64
544 #define NSI_ERR_MASK (NSI_FATAL_MASK | NSI_NON_FATAL_MASK)
545
546 static char *nsi_message[30] = {
547 "NSI Link Down", /* NSI_FERR/NSI_NERR bit 0, fatal error */
548 "", /* reserved */
549 "NSI Parity Error", /* bit 2, non-fatal */
550 "", /* reserved */
551 "", /* reserved */
552 "Correctable Error Message", /* bit 5, non-fatal */
553 "Non-Fatal Error Message", /* bit 6, non-fatal */
554 "Fatal Error Message", /* bit 7, fatal */
555 "", /* reserved */
556 "Receiver Error", /* bit 9, non-fatal */
557 "", /* reserved */
558 "Bad TLP", /* bit 11, non-fatal */
559 "Bad DLLP", /* bit 12, non-fatal */
560 "REPLAY_NUM Rollover", /* bit 13, non-fatal */
561 "", /* reserved */
562 "Replay Timer Timeout", /* bit 15, non-fatal */
563 "", /* reserved */
564 "", /* reserved */
565 "", /* reserved */
566 "Data Link Protocol Error", /* bit 19, fatal */
567 "", /* reserved */
568 "Poisoned TLP", /* bit 21, non-fatal */
569 "", /* reserved */
570 "Completion Timeout", /* bit 23, non-fatal */
571 "Completer Abort", /* bit 24, non-fatal */
572 "Unexpected Completion", /* bit 25, non-fatal */
573 "Receiver Overflow", /* bit 26, fatal */
574 "Malformed TLP", /* bit 27, fatal */
575 "", /* reserved */
576 "Unsupported Request" /* bit 29, non-fatal */
577 };
578
do_nsi_error(int fatal,u32 errors)579 static void do_nsi_error(int fatal, u32 errors)
580 {
581 int i;
582
583 for (i = 0; i < 30; i++) {
584 if (errors & (1 << i))
585 printk(KERN_WARNING "%sError %s\n",
586 fatal_message[fatal], nsi_message[i]);
587 }
588 }
589
nsi_error(int fatal,u32 errors,int * error_found,int handle_error)590 static inline void nsi_error(int fatal, u32 errors, int *error_found,
591 int handle_error)
592 {
593 *error_found = 1;
594
595 if (handle_error)
596 do_nsi_error(fatal, errors);
597 }
598
599 static char *membuf_message[4] = {
600 "Internal PMWB to DRAM parity",
601 "Internal PMWB to System Bus Parity",
602 "Internal System Bus or IO to PMWB Parity",
603 "Internal DRAM to PMWB Parity"
604 };
605
do_membuf_error(u8 errors)606 static void do_membuf_error(u8 errors)
607 {
608 int i;
609
610 for (i = 0; i < 4; i++) {
611 if (errors & (1 << i))
612 e752x_printk(KERN_WARNING, "Non-Fatal Error %s\n",
613 membuf_message[i]);
614 }
615 }
616
membuf_error(u8 errors,int * error_found,int handle_error)617 static inline void membuf_error(u8 errors, int *error_found, int handle_error)
618 {
619 *error_found = 1;
620
621 if (handle_error)
622 do_membuf_error(errors);
623 }
624
625 static char *sysbus_message[10] = {
626 "Addr or Request Parity",
627 "Data Strobe Glitch",
628 "Addr Strobe Glitch",
629 "Data Parity",
630 "Addr Above TOM",
631 "Non DRAM Lock Error",
632 "MCERR", "BINIT",
633 "Memory Parity",
634 "IO Subsystem Parity"
635 };
636
do_sysbus_error(int fatal,u32 errors)637 static void do_sysbus_error(int fatal, u32 errors)
638 {
639 int i;
640
641 for (i = 0; i < 10; i++) {
642 if (errors & (1 << i))
643 e752x_printk(KERN_WARNING, "%sError System Bus %s\n",
644 fatal_message[fatal], sysbus_message[i]);
645 }
646 }
647
sysbus_error(int fatal,u32 errors,int * error_found,int handle_error)648 static inline void sysbus_error(int fatal, u32 errors, int *error_found,
649 int handle_error)
650 {
651 *error_found = 1;
652
653 if (handle_error)
654 do_sysbus_error(fatal, errors);
655 }
656
e752x_check_hub_interface(struct e752x_error_info * info,int * error_found,int handle_error)657 static void e752x_check_hub_interface(struct e752x_error_info *info,
658 int *error_found, int handle_error)
659 {
660 u8 stat8;
661
662 //pci_read_config_byte(dev,E752X_HI_FERR,&stat8);
663
664 stat8 = info->hi_ferr;
665
666 if (stat8 & 0x7f) { /* Error, so process */
667 stat8 &= 0x7f;
668
669 if (stat8 & 0x2b)
670 hub_error(1, stat8 & 0x2b, error_found, handle_error);
671
672 if (stat8 & 0x54)
673 hub_error(0, stat8 & 0x54, error_found, handle_error);
674 }
675 //pci_read_config_byte(dev,E752X_HI_NERR,&stat8);
676
677 stat8 = info->hi_nerr;
678
679 if (stat8 & 0x7f) { /* Error, so process */
680 stat8 &= 0x7f;
681
682 if (stat8 & 0x2b)
683 hub_error(1, stat8 & 0x2b, error_found, handle_error);
684
685 if (stat8 & 0x54)
686 hub_error(0, stat8 & 0x54, error_found, handle_error);
687 }
688 }
689
e752x_check_ns_interface(struct e752x_error_info * info,int * error_found,int handle_error)690 static void e752x_check_ns_interface(struct e752x_error_info *info,
691 int *error_found, int handle_error)
692 {
693 u32 stat32;
694
695 stat32 = info->nsi_ferr;
696 if (stat32 & NSI_ERR_MASK) { /* Error, so process */
697 if (stat32 & NSI_FATAL_MASK) /* check for fatal errors */
698 nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
699 handle_error);
700 if (stat32 & NSI_NON_FATAL_MASK) /* check for non-fatal ones */
701 nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
702 handle_error);
703 }
704 stat32 = info->nsi_nerr;
705 if (stat32 & NSI_ERR_MASK) {
706 if (stat32 & NSI_FATAL_MASK)
707 nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
708 handle_error);
709 if (stat32 & NSI_NON_FATAL_MASK)
710 nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
711 handle_error);
712 }
713 }
714
e752x_check_sysbus(struct e752x_error_info * info,int * error_found,int handle_error)715 static void e752x_check_sysbus(struct e752x_error_info *info,
716 int *error_found, int handle_error)
717 {
718 u32 stat32, error32;
719
720 //pci_read_config_dword(dev,E752X_SYSBUS_FERR,&stat32);
721 stat32 = info->sysbus_ferr + (info->sysbus_nerr << 16);
722
723 if (stat32 == 0)
724 return; /* no errors */
725
726 error32 = (stat32 >> 16) & 0x3ff;
727 stat32 = stat32 & 0x3ff;
728
729 if (stat32 & 0x087)
730 sysbus_error(1, stat32 & 0x087, error_found, handle_error);
731
732 if (stat32 & 0x378)
733 sysbus_error(0, stat32 & 0x378, error_found, handle_error);
734
735 if (error32 & 0x087)
736 sysbus_error(1, error32 & 0x087, error_found, handle_error);
737
738 if (error32 & 0x378)
739 sysbus_error(0, error32 & 0x378, error_found, handle_error);
740 }
741
e752x_check_membuf(struct e752x_error_info * info,int * error_found,int handle_error)742 static void e752x_check_membuf(struct e752x_error_info *info,
743 int *error_found, int handle_error)
744 {
745 u8 stat8;
746
747 stat8 = info->buf_ferr;
748
749 if (stat8 & 0x0f) { /* Error, so process */
750 stat8 &= 0x0f;
751 membuf_error(stat8, error_found, handle_error);
752 }
753
754 stat8 = info->buf_nerr;
755
756 if (stat8 & 0x0f) { /* Error, so process */
757 stat8 &= 0x0f;
758 membuf_error(stat8, error_found, handle_error);
759 }
760 }
761
e752x_check_dram(struct mem_ctl_info * mci,struct e752x_error_info * info,int * error_found,int handle_error)762 static void e752x_check_dram(struct mem_ctl_info *mci,
763 struct e752x_error_info *info, int *error_found,
764 int handle_error)
765 {
766 u16 error_one, error_next;
767
768 error_one = info->dram_ferr;
769 error_next = info->dram_nerr;
770
771 /* decode and report errors */
772 if (error_one & 0x0101) /* check first error correctable */
773 process_ce(mci, error_one, info->dram_sec1_add,
774 info->dram_sec1_syndrome, error_found, handle_error);
775
776 if (error_next & 0x0101) /* check next error correctable */
777 process_ce(mci, error_next, info->dram_sec2_add,
778 info->dram_sec2_syndrome, error_found, handle_error);
779
780 if (error_one & 0x4040)
781 process_ue_no_info_wr(mci, error_found, handle_error);
782
783 if (error_next & 0x4040)
784 process_ue_no_info_wr(mci, error_found, handle_error);
785
786 if (error_one & 0x2020)
787 process_ded_retry(mci, error_one, info->dram_retr_add,
788 error_found, handle_error);
789
790 if (error_next & 0x2020)
791 process_ded_retry(mci, error_next, info->dram_retr_add,
792 error_found, handle_error);
793
794 if (error_one & 0x0808)
795 process_threshold_ce(mci, error_one, error_found, handle_error);
796
797 if (error_next & 0x0808)
798 process_threshold_ce(mci, error_next, error_found,
799 handle_error);
800
801 if (error_one & 0x0606)
802 process_ue(mci, error_one, info->dram_ded_add,
803 info->dram_scrb_add, error_found, handle_error);
804
805 if (error_next & 0x0606)
806 process_ue(mci, error_next, info->dram_ded_add,
807 info->dram_scrb_add, error_found, handle_error);
808 }
809
e752x_get_error_info(struct mem_ctl_info * mci,struct e752x_error_info * info)810 static void e752x_get_error_info(struct mem_ctl_info *mci,
811 struct e752x_error_info *info)
812 {
813 struct pci_dev *dev;
814 struct e752x_pvt *pvt;
815
816 memset(info, 0, sizeof(*info));
817 pvt = (struct e752x_pvt *)mci->pvt_info;
818 dev = pvt->dev_d0f1;
819 pci_read_config_dword(dev, E752X_FERR_GLOBAL, &info->ferr_global);
820
821 if (info->ferr_global) {
822 if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
823 pci_read_config_dword(dev, I3100_NSI_FERR,
824 &info->nsi_ferr);
825 info->hi_ferr = 0;
826 } else {
827 pci_read_config_byte(dev, E752X_HI_FERR,
828 &info->hi_ferr);
829 info->nsi_ferr = 0;
830 }
831 pci_read_config_word(dev, E752X_SYSBUS_FERR,
832 &info->sysbus_ferr);
833 pci_read_config_byte(dev, E752X_BUF_FERR, &info->buf_ferr);
834 pci_read_config_word(dev, E752X_DRAM_FERR, &info->dram_ferr);
835 pci_read_config_dword(dev, E752X_DRAM_SEC1_ADD,
836 &info->dram_sec1_add);
837 pci_read_config_word(dev, E752X_DRAM_SEC1_SYNDROME,
838 &info->dram_sec1_syndrome);
839 pci_read_config_dword(dev, E752X_DRAM_DED_ADD,
840 &info->dram_ded_add);
841 pci_read_config_dword(dev, E752X_DRAM_SCRB_ADD,
842 &info->dram_scrb_add);
843 pci_read_config_dword(dev, E752X_DRAM_RETR_ADD,
844 &info->dram_retr_add);
845
846 /* ignore the reserved bits just in case */
847 if (info->hi_ferr & 0x7f)
848 pci_write_config_byte(dev, E752X_HI_FERR,
849 info->hi_ferr);
850
851 if (info->nsi_ferr & NSI_ERR_MASK)
852 pci_write_config_dword(dev, I3100_NSI_FERR,
853 info->nsi_ferr);
854
855 if (info->sysbus_ferr)
856 pci_write_config_word(dev, E752X_SYSBUS_FERR,
857 info->sysbus_ferr);
858
859 if (info->buf_ferr & 0x0f)
860 pci_write_config_byte(dev, E752X_BUF_FERR,
861 info->buf_ferr);
862
863 if (info->dram_ferr)
864 pci_write_bits16(pvt->bridge_ck, E752X_DRAM_FERR,
865 info->dram_ferr, info->dram_ferr);
866
867 pci_write_config_dword(dev, E752X_FERR_GLOBAL,
868 info->ferr_global);
869 }
870
871 pci_read_config_dword(dev, E752X_NERR_GLOBAL, &info->nerr_global);
872
873 if (info->nerr_global) {
874 if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
875 pci_read_config_dword(dev, I3100_NSI_NERR,
876 &info->nsi_nerr);
877 info->hi_nerr = 0;
878 } else {
879 pci_read_config_byte(dev, E752X_HI_NERR,
880 &info->hi_nerr);
881 info->nsi_nerr = 0;
882 }
883 pci_read_config_word(dev, E752X_SYSBUS_NERR,
884 &info->sysbus_nerr);
885 pci_read_config_byte(dev, E752X_BUF_NERR, &info->buf_nerr);
886 pci_read_config_word(dev, E752X_DRAM_NERR, &info->dram_nerr);
887 pci_read_config_dword(dev, E752X_DRAM_SEC2_ADD,
888 &info->dram_sec2_add);
889 pci_read_config_word(dev, E752X_DRAM_SEC2_SYNDROME,
890 &info->dram_sec2_syndrome);
891
892 if (info->hi_nerr & 0x7f)
893 pci_write_config_byte(dev, E752X_HI_NERR,
894 info->hi_nerr);
895
896 if (info->nsi_nerr & NSI_ERR_MASK)
897 pci_write_config_dword(dev, I3100_NSI_NERR,
898 info->nsi_nerr);
899
900 if (info->sysbus_nerr)
901 pci_write_config_word(dev, E752X_SYSBUS_NERR,
902 info->sysbus_nerr);
903
904 if (info->buf_nerr & 0x0f)
905 pci_write_config_byte(dev, E752X_BUF_NERR,
906 info->buf_nerr);
907
908 if (info->dram_nerr)
909 pci_write_bits16(pvt->bridge_ck, E752X_DRAM_NERR,
910 info->dram_nerr, info->dram_nerr);
911
912 pci_write_config_dword(dev, E752X_NERR_GLOBAL,
913 info->nerr_global);
914 }
915 }
916
e752x_process_error_info(struct mem_ctl_info * mci,struct e752x_error_info * info,int handle_errors)917 static int e752x_process_error_info(struct mem_ctl_info *mci,
918 struct e752x_error_info *info,
919 int handle_errors)
920 {
921 u32 error32, stat32;
922 int error_found;
923
924 error_found = 0;
925 error32 = (info->ferr_global >> 18) & 0x3ff;
926 stat32 = (info->ferr_global >> 4) & 0x7ff;
927
928 if (error32)
929 global_error(1, error32, &error_found, handle_errors);
930
931 if (stat32)
932 global_error(0, stat32, &error_found, handle_errors);
933
934 error32 = (info->nerr_global >> 18) & 0x3ff;
935 stat32 = (info->nerr_global >> 4) & 0x7ff;
936
937 if (error32)
938 global_error(1, error32, &error_found, handle_errors);
939
940 if (stat32)
941 global_error(0, stat32, &error_found, handle_errors);
942
943 e752x_check_hub_interface(info, &error_found, handle_errors);
944 e752x_check_ns_interface(info, &error_found, handle_errors);
945 e752x_check_sysbus(info, &error_found, handle_errors);
946 e752x_check_membuf(info, &error_found, handle_errors);
947 e752x_check_dram(mci, info, &error_found, handle_errors);
948 return error_found;
949 }
950
e752x_check(struct mem_ctl_info * mci)951 static void e752x_check(struct mem_ctl_info *mci)
952 {
953 struct e752x_error_info info;
954
955 debugf3("%s()\n", __func__);
956 e752x_get_error_info(mci, &info);
957 e752x_process_error_info(mci, &info, 1);
958 }
959
960 /* Program byte/sec bandwidth scrub rate to hardware */
set_sdram_scrub_rate(struct mem_ctl_info * mci,u32 new_bw)961 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw)
962 {
963 const struct scrubrate *scrubrates;
964 struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
965 struct pci_dev *pdev = pvt->dev_d0f0;
966 int i;
967
968 if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
969 scrubrates = scrubrates_i3100;
970 else
971 scrubrates = scrubrates_e752x;
972
973 /* Translate the desired scrub rate to a e752x/3100 register value.
974 * Search for the bandwidth that is equal or greater than the
975 * desired rate and program the cooresponding register value.
976 */
977 for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
978 if (scrubrates[i].bandwidth >= new_bw)
979 break;
980
981 if (scrubrates[i].bandwidth == SDRATE_EOT)
982 return -1;
983
984 pci_write_config_word(pdev, E752X_MCHSCRB, scrubrates[i].scrubval);
985
986 return scrubrates[i].bandwidth;
987 }
988
989 /* Convert current scrub rate value into byte/sec bandwidth */
get_sdram_scrub_rate(struct mem_ctl_info * mci)990 static int get_sdram_scrub_rate(struct mem_ctl_info *mci)
991 {
992 const struct scrubrate *scrubrates;
993 struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
994 struct pci_dev *pdev = pvt->dev_d0f0;
995 u16 scrubval;
996 int i;
997
998 if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
999 scrubrates = scrubrates_i3100;
1000 else
1001 scrubrates = scrubrates_e752x;
1002
1003 /* Find the bandwidth matching the memory scrubber configuration */
1004 pci_read_config_word(pdev, E752X_MCHSCRB, &scrubval);
1005 scrubval = scrubval & 0x0f;
1006
1007 for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1008 if (scrubrates[i].scrubval == scrubval)
1009 break;
1010
1011 if (scrubrates[i].bandwidth == SDRATE_EOT) {
1012 e752x_printk(KERN_WARNING,
1013 "Invalid sdram scrub control value: 0x%x\n", scrubval);
1014 return -1;
1015 }
1016 return scrubrates[i].bandwidth;
1017
1018 }
1019
1020 /* Return 1 if dual channel mode is active. Else return 0. */
dual_channel_active(u16 ddrcsr)1021 static inline int dual_channel_active(u16 ddrcsr)
1022 {
1023 return (((ddrcsr >> 12) & 3) == 3);
1024 }
1025
1026 /* Remap csrow index numbers if map_type is "reverse"
1027 */
remap_csrow_index(struct mem_ctl_info * mci,int index)1028 static inline int remap_csrow_index(struct mem_ctl_info *mci, int index)
1029 {
1030 struct e752x_pvt *pvt = mci->pvt_info;
1031
1032 if (!pvt->map_type)
1033 return (7 - index);
1034
1035 return (index);
1036 }
1037
e752x_init_csrows(struct mem_ctl_info * mci,struct pci_dev * pdev,u16 ddrcsr)1038 static void e752x_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev,
1039 u16 ddrcsr)
1040 {
1041 struct csrow_info *csrow;
1042 unsigned long last_cumul_size;
1043 int index, mem_dev, drc_chan;
1044 int drc_drbg; /* DRB granularity 0=64mb, 1=128mb */
1045 int drc_ddim; /* DRAM Data Integrity Mode 0=none, 2=edac */
1046 u8 value;
1047 u32 dra, drc, cumul_size;
1048
1049 dra = 0;
1050 for (index = 0; index < 4; index++) {
1051 u8 dra_reg;
1052 pci_read_config_byte(pdev, E752X_DRA + index, &dra_reg);
1053 dra |= dra_reg << (index * 8);
1054 }
1055 pci_read_config_dword(pdev, E752X_DRC, &drc);
1056 drc_chan = dual_channel_active(ddrcsr);
1057 drc_drbg = drc_chan + 1; /* 128 in dual mode, 64 in single */
1058 drc_ddim = (drc >> 20) & 0x3;
1059
1060 /* The dram row boundary (DRB) reg values are boundary address for
1061 * each DRAM row with a granularity of 64 or 128MB (single/dual
1062 * channel operation). DRB regs are cumulative; therefore DRB7 will
1063 * contain the total memory contained in all eight rows.
1064 */
1065 for (last_cumul_size = index = 0; index < mci->nr_csrows; index++) {
1066 /* mem_dev 0=x8, 1=x4 */
1067 mem_dev = (dra >> (index * 4 + 2)) & 0x3;
1068 csrow = &mci->csrows[remap_csrow_index(mci, index)];
1069
1070 mem_dev = (mem_dev == 2);
1071 pci_read_config_byte(pdev, E752X_DRB + index, &value);
1072 /* convert a 128 or 64 MiB DRB to a page size. */
1073 cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
1074 debugf3("%s(): (%d) cumul_size 0x%x\n", __func__, index,
1075 cumul_size);
1076 if (cumul_size == last_cumul_size)
1077 continue; /* not populated */
1078
1079 csrow->first_page = last_cumul_size;
1080 csrow->last_page = cumul_size - 1;
1081 csrow->nr_pages = cumul_size - last_cumul_size;
1082 last_cumul_size = cumul_size;
1083 csrow->grain = 1 << 12; /* 4KiB - resolution of CELOG */
1084 csrow->mtype = MEM_RDDR; /* only one type supported */
1085 csrow->dtype = mem_dev ? DEV_X4 : DEV_X8;
1086
1087 /*
1088 * if single channel or x8 devices then SECDED
1089 * if dual channel and x4 then S4ECD4ED
1090 */
1091 if (drc_ddim) {
1092 if (drc_chan && mem_dev) {
1093 csrow->edac_mode = EDAC_S4ECD4ED;
1094 mci->edac_cap |= EDAC_FLAG_S4ECD4ED;
1095 } else {
1096 csrow->edac_mode = EDAC_SECDED;
1097 mci->edac_cap |= EDAC_FLAG_SECDED;
1098 }
1099 } else
1100 csrow->edac_mode = EDAC_NONE;
1101 }
1102 }
1103
e752x_init_mem_map_table(struct pci_dev * pdev,struct e752x_pvt * pvt)1104 static void e752x_init_mem_map_table(struct pci_dev *pdev,
1105 struct e752x_pvt *pvt)
1106 {
1107 int index;
1108 u8 value, last, row;
1109
1110 last = 0;
1111 row = 0;
1112
1113 for (index = 0; index < 8; index += 2) {
1114 pci_read_config_byte(pdev, E752X_DRB + index, &value);
1115 /* test if there is a dimm in this slot */
1116 if (value == last) {
1117 /* no dimm in the slot, so flag it as empty */
1118 pvt->map[index] = 0xff;
1119 pvt->map[index + 1] = 0xff;
1120 } else { /* there is a dimm in the slot */
1121 pvt->map[index] = row;
1122 row++;
1123 last = value;
1124 /* test the next value to see if the dimm is double
1125 * sided
1126 */
1127 pci_read_config_byte(pdev, E752X_DRB + index + 1,
1128 &value);
1129
1130 /* the dimm is single sided, so flag as empty */
1131 /* this is a double sided dimm to save the next row #*/
1132 pvt->map[index + 1] = (value == last) ? 0xff : row;
1133 row++;
1134 last = value;
1135 }
1136 }
1137 }
1138
1139 /* Return 0 on success or 1 on failure. */
e752x_get_devs(struct pci_dev * pdev,int dev_idx,struct e752x_pvt * pvt)1140 static int e752x_get_devs(struct pci_dev *pdev, int dev_idx,
1141 struct e752x_pvt *pvt)
1142 {
1143 struct pci_dev *dev;
1144
1145 pvt->bridge_ck = pci_get_device(PCI_VENDOR_ID_INTEL,
1146 pvt->dev_info->err_dev, pvt->bridge_ck);
1147
1148 if (pvt->bridge_ck == NULL)
1149 pvt->bridge_ck = pci_scan_single_device(pdev->bus,
1150 PCI_DEVFN(0, 1));
1151
1152 if (pvt->bridge_ck == NULL) {
1153 e752x_printk(KERN_ERR, "error reporting device not found:"
1154 "vendor %x device 0x%x (broken BIOS?)\n",
1155 PCI_VENDOR_ID_INTEL, e752x_devs[dev_idx].err_dev);
1156 return 1;
1157 }
1158
1159 dev = pci_get_device(PCI_VENDOR_ID_INTEL,
1160 e752x_devs[dev_idx].ctl_dev,
1161 NULL);
1162
1163 if (dev == NULL)
1164 goto fail;
1165
1166 pvt->dev_d0f0 = dev;
1167 pvt->dev_d0f1 = pci_dev_get(pvt->bridge_ck);
1168
1169 return 0;
1170
1171 fail:
1172 pci_dev_put(pvt->bridge_ck);
1173 return 1;
1174 }
1175
1176 /* Setup system bus parity mask register.
1177 * Sysbus parity supported on:
1178 * e7320/e7520/e7525 + Xeon
1179 */
e752x_init_sysbus_parity_mask(struct e752x_pvt * pvt)1180 static void e752x_init_sysbus_parity_mask(struct e752x_pvt *pvt)
1181 {
1182 char *cpu_id = cpu_data(0).x86_model_id;
1183 struct pci_dev *dev = pvt->dev_d0f1;
1184 int enable = 1;
1185
1186 /* Allow module parameter override, else see if CPU supports parity */
1187 if (sysbus_parity != -1) {
1188 enable = sysbus_parity;
1189 } else if (cpu_id[0] && !strstr(cpu_id, "Xeon")) {
1190 e752x_printk(KERN_INFO, "System Bus Parity not "
1191 "supported by CPU, disabling\n");
1192 enable = 0;
1193 }
1194
1195 if (enable)
1196 pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0000);
1197 else
1198 pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0309);
1199 }
1200
e752x_init_error_reporting_regs(struct e752x_pvt * pvt)1201 static void e752x_init_error_reporting_regs(struct e752x_pvt *pvt)
1202 {
1203 struct pci_dev *dev;
1204
1205 dev = pvt->dev_d0f1;
1206 /* Turn off error disable & SMI in case the BIOS turned it on */
1207 if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
1208 pci_write_config_dword(dev, I3100_NSI_EMASK, 0);
1209 pci_write_config_dword(dev, I3100_NSI_SMICMD, 0);
1210 } else {
1211 pci_write_config_byte(dev, E752X_HI_ERRMASK, 0x00);
1212 pci_write_config_byte(dev, E752X_HI_SMICMD, 0x00);
1213 }
1214
1215 e752x_init_sysbus_parity_mask(pvt);
1216
1217 pci_write_config_word(dev, E752X_SYSBUS_SMICMD, 0x00);
1218 pci_write_config_byte(dev, E752X_BUF_ERRMASK, 0x00);
1219 pci_write_config_byte(dev, E752X_BUF_SMICMD, 0x00);
1220 pci_write_config_byte(dev, E752X_DRAM_ERRMASK, 0x00);
1221 pci_write_config_byte(dev, E752X_DRAM_SMICMD, 0x00);
1222 }
1223
e752x_probe1(struct pci_dev * pdev,int dev_idx)1224 static int e752x_probe1(struct pci_dev *pdev, int dev_idx)
1225 {
1226 u16 pci_data;
1227 u8 stat8;
1228 struct mem_ctl_info *mci;
1229 struct e752x_pvt *pvt;
1230 u16 ddrcsr;
1231 int drc_chan; /* Number of channels 0=1chan,1=2chan */
1232 struct e752x_error_info discard;
1233
1234 debugf0("%s(): mci\n", __func__);
1235 debugf0("Starting Probe1\n");
1236
1237 /* check to see if device 0 function 1 is enabled; if it isn't, we
1238 * assume the BIOS has reserved it for a reason and is expecting
1239 * exclusive access, we take care not to violate that assumption and
1240 * fail the probe. */
1241 pci_read_config_byte(pdev, E752X_DEVPRES1, &stat8);
1242 if (!force_function_unhide && !(stat8 & (1 << 5))) {
1243 printk(KERN_INFO "Contact your BIOS vendor to see if the "
1244 "E752x error registers can be safely un-hidden\n");
1245 return -ENODEV;
1246 }
1247 stat8 |= (1 << 5);
1248 pci_write_config_byte(pdev, E752X_DEVPRES1, stat8);
1249
1250 pci_read_config_word(pdev, E752X_DDRCSR, &ddrcsr);
1251 /* FIXME: should check >>12 or 0xf, true for all? */
1252 /* Dual channel = 1, Single channel = 0 */
1253 drc_chan = dual_channel_active(ddrcsr);
1254
1255 mci = edac_mc_alloc(sizeof(*pvt), E752X_NR_CSROWS, drc_chan + 1, 0);
1256
1257 if (mci == NULL) {
1258 return -ENOMEM;
1259 }
1260
1261 debugf3("%s(): init mci\n", __func__);
1262 mci->mtype_cap = MEM_FLAG_RDDR;
1263 /* 3100 IMCH supports SECDEC only */
1264 mci->edac_ctl_cap = (dev_idx == I3100) ? EDAC_FLAG_SECDED :
1265 (EDAC_FLAG_NONE | EDAC_FLAG_SECDED | EDAC_FLAG_S4ECD4ED);
1266 /* FIXME - what if different memory types are in different csrows? */
1267 mci->mod_name = EDAC_MOD_STR;
1268 mci->mod_ver = E752X_REVISION;
1269 mci->dev = &pdev->dev;
1270
1271 debugf3("%s(): init pvt\n", __func__);
1272 pvt = (struct e752x_pvt *)mci->pvt_info;
1273 pvt->dev_info = &e752x_devs[dev_idx];
1274 pvt->mc_symmetric = ((ddrcsr & 0x10) != 0);
1275
1276 if (e752x_get_devs(pdev, dev_idx, pvt)) {
1277 edac_mc_free(mci);
1278 return -ENODEV;
1279 }
1280
1281 debugf3("%s(): more mci init\n", __func__);
1282 mci->ctl_name = pvt->dev_info->ctl_name;
1283 mci->dev_name = pci_name(pdev);
1284 mci->edac_check = e752x_check;
1285 mci->ctl_page_to_phys = ctl_page_to_phys;
1286 mci->set_sdram_scrub_rate = set_sdram_scrub_rate;
1287 mci->get_sdram_scrub_rate = get_sdram_scrub_rate;
1288
1289 /* set the map type. 1 = normal, 0 = reversed
1290 * Must be set before e752x_init_csrows in case csrow mapping
1291 * is reversed.
1292 */
1293 pci_read_config_byte(pdev, E752X_DRM, &stat8);
1294 pvt->map_type = ((stat8 & 0x0f) > ((stat8 >> 4) & 0x0f));
1295
1296 e752x_init_csrows(mci, pdev, ddrcsr);
1297 e752x_init_mem_map_table(pdev, pvt);
1298
1299 if (dev_idx == I3100)
1300 mci->edac_cap = EDAC_FLAG_SECDED; /* the only mode supported */
1301 else
1302 mci->edac_cap |= EDAC_FLAG_NONE;
1303 debugf3("%s(): tolm, remapbase, remaplimit\n", __func__);
1304
1305 /* load the top of low memory, remap base, and remap limit vars */
1306 pci_read_config_word(pdev, E752X_TOLM, &pci_data);
1307 pvt->tolm = ((u32) pci_data) << 4;
1308 pci_read_config_word(pdev, E752X_REMAPBASE, &pci_data);
1309 pvt->remapbase = ((u32) pci_data) << 14;
1310 pci_read_config_word(pdev, E752X_REMAPLIMIT, &pci_data);
1311 pvt->remaplimit = ((u32) pci_data) << 14;
1312 e752x_printk(KERN_INFO,
1313 "tolm = %x, remapbase = %x, remaplimit = %x\n",
1314 pvt->tolm, pvt->remapbase, pvt->remaplimit);
1315
1316 /* Here we assume that we will never see multiple instances of this
1317 * type of memory controller. The ID is therefore hardcoded to 0.
1318 */
1319 if (edac_mc_add_mc(mci)) {
1320 debugf3("%s(): failed edac_mc_add_mc()\n", __func__);
1321 goto fail;
1322 }
1323
1324 e752x_init_error_reporting_regs(pvt);
1325 e752x_get_error_info(mci, &discard); /* clear other MCH errors */
1326
1327 /* allocating generic PCI control info */
1328 e752x_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR);
1329 if (!e752x_pci) {
1330 printk(KERN_WARNING
1331 "%s(): Unable to create PCI control\n", __func__);
1332 printk(KERN_WARNING
1333 "%s(): PCI error report via EDAC not setup\n",
1334 __func__);
1335 }
1336
1337 /* get this far and it's successful */
1338 debugf3("%s(): success\n", __func__);
1339 return 0;
1340
1341 fail:
1342 pci_dev_put(pvt->dev_d0f0);
1343 pci_dev_put(pvt->dev_d0f1);
1344 pci_dev_put(pvt->bridge_ck);
1345 edac_mc_free(mci);
1346
1347 return -ENODEV;
1348 }
1349
1350 /* returns count (>= 0), or negative on error */
e752x_init_one(struct pci_dev * pdev,const struct pci_device_id * ent)1351 static int __devinit e752x_init_one(struct pci_dev *pdev,
1352 const struct pci_device_id *ent)
1353 {
1354 debugf0("%s()\n", __func__);
1355
1356 /* wake up and enable device */
1357 if (pci_enable_device(pdev) < 0)
1358 return -EIO;
1359
1360 return e752x_probe1(pdev, ent->driver_data);
1361 }
1362
e752x_remove_one(struct pci_dev * pdev)1363 static void __devexit e752x_remove_one(struct pci_dev *pdev)
1364 {
1365 struct mem_ctl_info *mci;
1366 struct e752x_pvt *pvt;
1367
1368 debugf0("%s()\n", __func__);
1369
1370 if (e752x_pci)
1371 edac_pci_release_generic_ctl(e752x_pci);
1372
1373 if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL)
1374 return;
1375
1376 pvt = (struct e752x_pvt *)mci->pvt_info;
1377 pci_dev_put(pvt->dev_d0f0);
1378 pci_dev_put(pvt->dev_d0f1);
1379 pci_dev_put(pvt->bridge_ck);
1380 edac_mc_free(mci);
1381 }
1382
1383 static const struct pci_device_id e752x_pci_tbl[] __devinitdata = {
1384 {
1385 PCI_VEND_DEV(INTEL, 7520_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1386 E7520},
1387 {
1388 PCI_VEND_DEV(INTEL, 7525_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1389 E7525},
1390 {
1391 PCI_VEND_DEV(INTEL, 7320_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1392 E7320},
1393 {
1394 PCI_VEND_DEV(INTEL, 3100_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1395 I3100},
1396 {
1397 0,
1398 } /* 0 terminated list. */
1399 };
1400
1401 MODULE_DEVICE_TABLE(pci, e752x_pci_tbl);
1402
1403 static struct pci_driver e752x_driver = {
1404 .name = EDAC_MOD_STR,
1405 .probe = e752x_init_one,
1406 .remove = __devexit_p(e752x_remove_one),
1407 .id_table = e752x_pci_tbl,
1408 };
1409
e752x_init(void)1410 static int __init e752x_init(void)
1411 {
1412 int pci_rc;
1413
1414 debugf3("%s()\n", __func__);
1415
1416 /* Ensure that the OPSTATE is set correctly for POLL or NMI */
1417 opstate_init();
1418
1419 pci_rc = pci_register_driver(&e752x_driver);
1420 return (pci_rc < 0) ? pci_rc : 0;
1421 }
1422
e752x_exit(void)1423 static void __exit e752x_exit(void)
1424 {
1425 debugf3("%s()\n", __func__);
1426 pci_unregister_driver(&e752x_driver);
1427 }
1428
1429 module_init(e752x_init);
1430 module_exit(e752x_exit);
1431
1432 MODULE_LICENSE("GPL");
1433 MODULE_AUTHOR("Linux Networx (http://lnxi.com) Tom Zimmerman\n");
1434 MODULE_DESCRIPTION("MC support for Intel e752x/3100 memory controllers");
1435
1436 module_param(force_function_unhide, int, 0444);
1437 MODULE_PARM_DESC(force_function_unhide, "if BIOS sets Dev0:Fun1 up as hidden:"
1438 " 1=force unhide and hope BIOS doesn't fight driver for "
1439 "Dev0:Fun1 access");
1440
1441 module_param(edac_op_state, int, 0444);
1442 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
1443
1444 module_param(sysbus_parity, int, 0444);
1445 MODULE_PARM_DESC(sysbus_parity, "0=disable system bus parity checking,"
1446 " 1=enable system bus parity checking, default=auto-detect");
1447 module_param(report_non_memory_errors, int, 0644);
1448 MODULE_PARM_DESC(report_non_memory_errors, "0=disable non-memory error "
1449 "reporting, 1=enable non-memory error reporting");
1450