Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
e752x_edac.c
Go to the documentation of this file.
1 /*
2  * Intel e752x Memory Controller kernel module
3  * (C) 2004 Linux Networx (http://lnxi.com)
4  * This file may be distributed under the terms of the
5  * GNU General Public License.
6  *
7  * Implement support for the e7520, E7525, e7320 and i3100 memory controllers.
8  *
9  * Datasheets:
10  * http://www.intel.in/content/www/in/en/chipsets/e7525-memory-controller-hub-datasheet.html
11  * ftp://download.intel.com/design/intarch/datashts/31345803.pdf
12  *
13  * Written by Tom Zimmerman
14  *
15  * Contributors:
16  * Thayne Harbaugh at realmsys.com (?)
17  * Wang Zhenyu at intel.com
18  * Dave Jiang at mvista.com
19  *
20  */
21 
22 #include <linux/module.h>
23 #include <linux/init.h>
24 #include <linux/pci.h>
25 #include <linux/pci_ids.h>
26 #include <linux/edac.h>
27 #include "edac_core.h"
28 
29 #define E752X_REVISION " Ver: 2.0.2"
30 #define EDAC_MOD_STR "e752x_edac"
31 
32 static int report_non_memory_errors;
33 static int force_function_unhide;
34 static int sysbus_parity = -1;
35 
36 static struct edac_pci_ctl_info *e752x_pci;
37 
38 #define e752x_printk(level, fmt, arg...) \
39  edac_printk(level, "e752x", fmt, ##arg)
40 
41 #define e752x_mc_printk(mci, level, fmt, arg...) \
42  edac_mc_chipset_printk(mci, level, "e752x", fmt, ##arg)
43 
44 #ifndef PCI_DEVICE_ID_INTEL_7520_0
45 #define PCI_DEVICE_ID_INTEL_7520_0 0x3590
46 #endif /* PCI_DEVICE_ID_INTEL_7520_0 */
47 
48 #ifndef PCI_DEVICE_ID_INTEL_7520_1_ERR
49 #define PCI_DEVICE_ID_INTEL_7520_1_ERR 0x3591
50 #endif /* PCI_DEVICE_ID_INTEL_7520_1_ERR */
51 
52 #ifndef PCI_DEVICE_ID_INTEL_7525_0
53 #define PCI_DEVICE_ID_INTEL_7525_0 0x359E
54 #endif /* PCI_DEVICE_ID_INTEL_7525_0 */
55 
56 #ifndef PCI_DEVICE_ID_INTEL_7525_1_ERR
57 #define PCI_DEVICE_ID_INTEL_7525_1_ERR 0x3593
58 #endif /* PCI_DEVICE_ID_INTEL_7525_1_ERR */
59 
60 #ifndef PCI_DEVICE_ID_INTEL_7320_0
61 #define PCI_DEVICE_ID_INTEL_7320_0 0x3592
62 #endif /* PCI_DEVICE_ID_INTEL_7320_0 */
63 
64 #ifndef PCI_DEVICE_ID_INTEL_7320_1_ERR
65 #define PCI_DEVICE_ID_INTEL_7320_1_ERR 0x3593
66 #endif /* PCI_DEVICE_ID_INTEL_7320_1_ERR */
67 
68 #ifndef PCI_DEVICE_ID_INTEL_3100_0
69 #define PCI_DEVICE_ID_INTEL_3100_0 0x35B0
70 #endif /* PCI_DEVICE_ID_INTEL_3100_0 */
71 
72 #ifndef PCI_DEVICE_ID_INTEL_3100_1_ERR
73 #define PCI_DEVICE_ID_INTEL_3100_1_ERR 0x35B1
74 #endif /* PCI_DEVICE_ID_INTEL_3100_1_ERR */
75 
76 #define E752X_NR_CSROWS 8 /* number of csrows */
77 
78 /* E752X register addresses - device 0 function 0 */
79 #define E752X_MCHSCRB 0x52 /* Memory Scrub register (16b) */
80  /*
81  * 6:5 Scrub Completion Count
82  * 3:2 Scrub Rate (i3100 only)
83  * 01=fast 10=normal
84  * 1:0 Scrub Mode enable
85  * 00=off 10=on
86  */
87 #define E752X_DRB 0x60 /* DRAM row boundary register (8b) */
88 #define E752X_DRA 0x70 /* DRAM row attribute register (8b) */
89  /*
90  * 31:30 Device width row 7
91  * 01=x8 10=x4 11=x8 DDR2
92  * 27:26 Device width row 6
93  * 23:22 Device width row 5
94  * 19:20 Device width row 4
95  * 15:14 Device width row 3
96  * 11:10 Device width row 2
97  * 7:6 Device width row 1
98  * 3:2 Device width row 0
99  */
100 #define E752X_DRC 0x7C /* DRAM controller mode reg (32b) */
101  /* FIXME:IS THIS RIGHT? */
102  /*
103  * 22 Number channels 0=1,1=2
104  * 19:18 DRB Granularity 32/64MB
105  */
106 #define E752X_DRM 0x80 /* Dimm mapping register */
107 #define E752X_DDRCSR 0x9A /* DDR control and status reg (16b) */
108  /*
109  * 14:12 1 single A, 2 single B, 3 dual
110  */
111 #define E752X_TOLM 0xC4 /* DRAM top of low memory reg (16b) */
112 #define E752X_REMAPBASE 0xC6 /* DRAM remap base address reg (16b) */
113 #define E752X_REMAPLIMIT 0xC8 /* DRAM remap limit address reg (16b) */
114 #define E752X_REMAPOFFSET 0xCA /* DRAM remap limit offset reg (16b) */
115 
116 /* E752X register addresses - device 0 function 1 */
117 #define E752X_FERR_GLOBAL 0x40 /* Global first error register (32b) */
118 #define E752X_NERR_GLOBAL 0x44 /* Global next error register (32b) */
119 #define E752X_HI_FERR 0x50 /* Hub interface first error reg (8b) */
120 #define E752X_HI_NERR 0x52 /* Hub interface next error reg (8b) */
121 #define E752X_HI_ERRMASK 0x54 /* Hub interface error mask reg (8b) */
122 #define E752X_HI_SMICMD 0x5A /* Hub interface SMI command reg (8b) */
123 #define E752X_SYSBUS_FERR 0x60 /* System buss first error reg (16b) */
124 #define E752X_SYSBUS_NERR 0x62 /* System buss next error reg (16b) */
125 #define E752X_SYSBUS_ERRMASK 0x64 /* System buss error mask reg (16b) */
126 #define E752X_SYSBUS_SMICMD 0x6A /* System buss SMI command reg (16b) */
127 #define E752X_BUF_FERR 0x70 /* Memory buffer first error reg (8b) */
128 #define E752X_BUF_NERR 0x72 /* Memory buffer next error reg (8b) */
129 #define E752X_BUF_ERRMASK 0x74 /* Memory buffer error mask reg (8b) */
130 #define E752X_BUF_SMICMD 0x7A /* Memory buffer SMI cmd reg (8b) */
131 #define E752X_DRAM_FERR 0x80 /* DRAM first error register (16b) */
132 #define E752X_DRAM_NERR 0x82 /* DRAM next error register (16b) */
133 #define E752X_DRAM_ERRMASK 0x84 /* DRAM error mask register (8b) */
134 #define E752X_DRAM_SMICMD 0x8A /* DRAM SMI command register (8b) */
135 #define E752X_DRAM_RETR_ADD 0xAC /* DRAM Retry address register (32b) */
136 #define E752X_DRAM_SEC1_ADD 0xA0 /* DRAM first correctable memory */
137  /* error address register (32b) */
138  /*
139  * 31 Reserved
140  * 30:2 CE address (64 byte block 34:6
141  * 1 Reserved
142  * 0 HiLoCS
143  */
144 #define E752X_DRAM_SEC2_ADD 0xC8 /* DRAM first correctable memory */
145  /* error address register (32b) */
146  /*
147  * 31 Reserved
148  * 30:2 CE address (64 byte block 34:6)
149  * 1 Reserved
150  * 0 HiLoCS
151  */
152 #define E752X_DRAM_DED_ADD 0xA4 /* DRAM first uncorrectable memory */
153  /* error address register (32b) */
154  /*
155  * 31 Reserved
156  * 30:2 CE address (64 byte block 34:6)
157  * 1 Reserved
158  * 0 HiLoCS
159  */
160 #define E752X_DRAM_SCRB_ADD 0xA8 /* DRAM 1st uncorrectable scrub mem */
161  /* error address register (32b) */
162  /*
163  * 31 Reserved
164  * 30:2 CE address (64 byte block 34:6
165  * 1 Reserved
166  * 0 HiLoCS
167  */
168 #define E752X_DRAM_SEC1_SYNDROME 0xC4 /* DRAM first correctable memory */
169  /* error syndrome register (16b) */
170 #define E752X_DRAM_SEC2_SYNDROME 0xC6 /* DRAM second correctable memory */
171  /* error syndrome register (16b) */
172 #define E752X_DEVPRES1 0xF4 /* Device Present 1 register (8b) */
173 
174 /* 3100 IMCH specific register addresses - device 0 function 1 */
175 #define I3100_NSI_FERR 0x48 /* NSI first error reg (32b) */
176 #define I3100_NSI_NERR 0x4C /* NSI next error reg (32b) */
177 #define I3100_NSI_SMICMD 0x54 /* NSI SMI command register (32b) */
178 #define I3100_NSI_EMASK 0x90 /* NSI error mask register (32b) */
179 
180 /* ICH5R register addresses - device 30 function 0 */
181 #define ICH5R_PCI_STAT 0x06 /* PCI status register (16b) */
182 #define ICH5R_PCI_2ND_STAT 0x1E /* PCI status secondary reg (16b) */
183 #define ICH5R_PCI_BRIDGE_CTL 0x3E /* PCI bridge control register (16b) */
184 
186  E7520 = 0,
187  E7525 = 1,
188  E7320 = 2,
189  I3100 = 3
190 };
191 
192 /*
193  * Those chips Support single-rank and dual-rank memories only.
194  *
195  * On e752x chips, the odd rows are present only on dual-rank memories.
196  * Dividing the rank by two will provide the dimm#
197  *
198  * i3100 MC has a different mapping: it supports only 4 ranks.
199  *
200  * The mapping is (from 1 to n):
201  * slot single-ranked double-ranked
202  * dimm #1 -> rank #4 NA
203  * dimm #2 -> rank #3 NA
204  * dimm #3 -> rank #2 Ranks 2 and 3
205  * dimm #4 -> rank $1 Ranks 1 and 4
206  *
207  * FIXME: The current mapping for i3100 considers that it supports up to 8
208  * ranks/chanel, but datasheet says that the MC supports only 4 ranks.
209  */
210 
211 struct e752x_pvt {
213  struct pci_dev *dev_d0f0;
214  struct pci_dev *dev_d0f1;
219  u8 map[8];
220  int map_type;
221  const struct e752x_dev_info *dev_info;
222 };
223 
227  const char *ctl_name;
228 };
229 
233  u32 nsi_ferr; /* 3100 only */
234  u32 nsi_nerr; /* 3100 only */
235  u8 hi_ferr; /* all but 3100 */
236  u8 hi_nerr; /* all but 3100 */
250 };
251 
252 static const struct e752x_dev_info e752x_devs[] = {
253  [E7520] = {
255  .ctl_dev = PCI_DEVICE_ID_INTEL_7520_0,
256  .ctl_name = "E7520"},
257  [E7525] = {
259  .ctl_dev = PCI_DEVICE_ID_INTEL_7525_0,
260  .ctl_name = "E7525"},
261  [E7320] = {
263  .ctl_dev = PCI_DEVICE_ID_INTEL_7320_0,
264  .ctl_name = "E7320"},
265  [I3100] = {
267  .ctl_dev = PCI_DEVICE_ID_INTEL_3100_0,
268  .ctl_name = "3100"},
269 };
270 
271 /* Valid scrub rates for the e752x/3100 hardware memory scrubber. We
272  * map the scrubbing bandwidth to a hardware register value. The 'set'
273  * operation finds the 'matching or higher value'. Note that scrubbing
274  * on the e752x can only be enabled/disabled. The 3100 supports
275  * a normal and fast mode.
276  */
277 
278 #define SDRATE_EOT 0xFFFFFFFF
279 
280 struct scrubrate {
281  u32 bandwidth; /* bandwidth consumed by scrubbing in bytes/sec */
282  u16 scrubval; /* register value for scrub rate */
283 };
284 
285 /* Rate below assumes same performance as i3100 using PC3200 DDR2 in
286  * normal mode. e752x bridges don't support choosing normal or fast mode,
287  * so the scrubbing bandwidth value isn't all that important - scrubbing is
288  * either on or off.
289  */
290 static const struct scrubrate scrubrates_e752x[] = {
291  {0, 0x00}, /* Scrubbing Off */
292  {500000, 0x02}, /* Scrubbing On */
293  {SDRATE_EOT, 0x00} /* End of Table */
294 };
295 
296 /* Fast mode: 2 GByte PC3200 DDR2 scrubbed in 33s = 63161283 bytes/s
297  * Normal mode: 125 (32000 / 256) times slower than fast mode.
298  */
299 static const struct scrubrate scrubrates_i3100[] = {
300  {0, 0x00}, /* Scrubbing Off */
301  {500000, 0x0a}, /* Normal mode - 32k clocks */
302  {62500000, 0x06}, /* Fast mode - 256 clocks */
303  {SDRATE_EOT, 0x00} /* End of Table */
304 };
305 
306 static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci,
307  unsigned long page)
308 {
309  u32 remap;
310  struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
311 
312  edac_dbg(3, "\n");
313 
314  if (page < pvt->tolm)
315  return page;
316 
317  if ((page >= 0x100000) && (page < pvt->remapbase))
318  return page;
319 
320  remap = (page - pvt->tolm) + pvt->remapbase;
321 
322  if (remap < pvt->remaplimit)
323  return remap;
324 
325  e752x_printk(KERN_ERR, "Invalid page %lx - out of range\n", page);
326  return pvt->tolm - 1;
327 }
328 
329 static void do_process_ce(struct mem_ctl_info *mci, u16 error_one,
330  u32 sec1_add, u16 sec1_syndrome)
331 {
332  u32 page;
333  int row;
334  int channel;
335  int i;
336  struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
337 
338  edac_dbg(3, "\n");
339 
340  /* convert the addr to 4k page */
341  page = sec1_add >> (PAGE_SHIFT - 4);
342 
343  /* FIXME - check for -1 */
344  if (pvt->mc_symmetric) {
345  /* chip select are bits 14 & 13 */
346  row = ((page >> 1) & 3);
348  "Test row %d Table %d %d %d %d %d %d %d %d\n", row,
349  pvt->map[0], pvt->map[1], pvt->map[2], pvt->map[3],
350  pvt->map[4], pvt->map[5], pvt->map[6],
351  pvt->map[7]);
352 
353  /* test for channel remapping */
354  for (i = 0; i < 8; i++) {
355  if (pvt->map[i] == row)
356  break;
357  }
358 
359  e752x_printk(KERN_WARNING, "Test computed row %d\n", i);
360 
361  if (i < 8)
362  row = i;
363  else
365  "row %d not found in remap table\n",
366  row);
367  } else
368  row = edac_mc_find_csrow_by_page(mci, page);
369 
370  /* 0 = channel A, 1 = channel B */
371  channel = !(error_one & 1);
372 
373  /* e752x mc reads 34:6 of the DRAM linear address */
375  page, offset_in_page(sec1_add << 4), sec1_syndrome,
376  row, channel, -1,
377  "e752x CE", "");
378 }
379 
380 static inline void process_ce(struct mem_ctl_info *mci, u16 error_one,
381  u32 sec1_add, u16 sec1_syndrome, int *error_found,
382  int handle_error)
383 {
384  *error_found = 1;
385 
386  if (handle_error)
387  do_process_ce(mci, error_one, sec1_add, sec1_syndrome);
388 }
389 
390 static void do_process_ue(struct mem_ctl_info *mci, u16 error_one,
391  u32 ded_add, u32 scrb_add)
392 {
393  u32 error_2b, block_page;
394  int row;
395  struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
396 
397  edac_dbg(3, "\n");
398 
399  if (error_one & 0x0202) {
400  error_2b = ded_add;
401 
402  /* convert to 4k address */
403  block_page = error_2b >> (PAGE_SHIFT - 4);
404 
405  row = pvt->mc_symmetric ?
406  /* chip select are bits 14 & 13 */
407  ((block_page >> 1) & 3) :
408  edac_mc_find_csrow_by_page(mci, block_page);
409 
410  /* e752x mc reads 34:6 of the DRAM linear address */
412  block_page,
413  offset_in_page(error_2b << 4), 0,
414  row, -1, -1,
415  "e752x UE from Read", "");
416 
417  }
418  if (error_one & 0x0404) {
419  error_2b = scrb_add;
420 
421  /* convert to 4k address */
422  block_page = error_2b >> (PAGE_SHIFT - 4);
423 
424  row = pvt->mc_symmetric ?
425  /* chip select are bits 14 & 13 */
426  ((block_page >> 1) & 3) :
427  edac_mc_find_csrow_by_page(mci, block_page);
428 
429  /* e752x mc reads 34:6 of the DRAM linear address */
431  block_page,
432  offset_in_page(error_2b << 4), 0,
433  row, -1, -1,
434  "e752x UE from Scruber", "");
435  }
436 }
437 
438 static inline void process_ue(struct mem_ctl_info *mci, u16 error_one,
439  u32 ded_add, u32 scrb_add, int *error_found,
440  int handle_error)
441 {
442  *error_found = 1;
443 
444  if (handle_error)
445  do_process_ue(mci, error_one, ded_add, scrb_add);
446 }
447 
448 static inline void process_ue_no_info_wr(struct mem_ctl_info *mci,
449  int *error_found, int handle_error)
450 {
451  *error_found = 1;
452 
453  if (!handle_error)
454  return;
455 
456  edac_dbg(3, "\n");
458  -1, -1, -1,
459  "e752x UE log memory write", "");
460 }
461 
462 static void do_process_ded_retry(struct mem_ctl_info *mci, u16 error,
463  u32 retry_add)
464 {
465  u32 error_1b, page;
466  int row;
467  struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
468 
469  error_1b = retry_add;
470  page = error_1b >> (PAGE_SHIFT - 4); /* convert the addr to 4k page */
471 
472  /* chip select are bits 14 & 13 */
473  row = pvt->mc_symmetric ? ((page >> 1) & 3) :
474  edac_mc_find_csrow_by_page(mci, page);
475 
477  "CE page 0x%lx, row %d : Memory read retry\n",
478  (long unsigned int)page, row);
479 }
480 
481 static inline void process_ded_retry(struct mem_ctl_info *mci, u16 error,
482  u32 retry_add, int *error_found,
483  int handle_error)
484 {
485  *error_found = 1;
486 
487  if (handle_error)
488  do_process_ded_retry(mci, error, retry_add);
489 }
490 
491 static inline void process_threshold_ce(struct mem_ctl_info *mci, u16 error,
492  int *error_found, int handle_error)
493 {
494  *error_found = 1;
495 
496  if (handle_error)
497  e752x_mc_printk(mci, KERN_WARNING, "Memory threshold CE\n");
498 }
499 
500 static char *global_message[11] = {
501  "PCI Express C1",
502  "PCI Express C",
503  "PCI Express B1",
504  "PCI Express B",
505  "PCI Express A1",
506  "PCI Express A",
507  "DMA Controller",
508  "HUB or NS Interface",
509  "System Bus",
510  "DRAM Controller", /* 9th entry */
511  "Internal Buffer"
512 };
513 
514 #define DRAM_ENTRY 9
515 
516 static char *fatal_message[2] = { "Non-Fatal ", "Fatal " };
517 
518 static void do_global_error(int fatal, u32 errors)
519 {
520  int i;
521 
522  for (i = 0; i < 11; i++) {
523  if (errors & (1 << i)) {
524  /* If the error is from DRAM Controller OR
525  * we are to report ALL errors, then
526  * report the error
527  */
528  if ((i == DRAM_ENTRY) || report_non_memory_errors)
529  e752x_printk(KERN_WARNING, "%sError %s\n",
530  fatal_message[fatal],
531  global_message[i]);
532  }
533  }
534 }
535 
536 static inline void global_error(int fatal, u32 errors, int *error_found,
537  int handle_error)
538 {
539  *error_found = 1;
540 
541  if (handle_error)
542  do_global_error(fatal, errors);
543 }
544 
545 static char *hub_message[7] = {
546  "HI Address or Command Parity", "HI Illegal Access",
547  "HI Internal Parity", "Out of Range Access",
548  "HI Data Parity", "Enhanced Config Access",
549  "Hub Interface Target Abort"
550 };
551 
552 static void do_hub_error(int fatal, u8 errors)
553 {
554  int i;
555 
556  for (i = 0; i < 7; i++) {
557  if (errors & (1 << i))
558  e752x_printk(KERN_WARNING, "%sError %s\n",
559  fatal_message[fatal], hub_message[i]);
560  }
561 }
562 
563 static inline void hub_error(int fatal, u8 errors, int *error_found,
564  int handle_error)
565 {
566  *error_found = 1;
567 
568  if (handle_error)
569  do_hub_error(fatal, errors);
570 }
571 
572 #define NSI_FATAL_MASK 0x0c080081
573 #define NSI_NON_FATAL_MASK 0x23a0ba64
574 #define NSI_ERR_MASK (NSI_FATAL_MASK | NSI_NON_FATAL_MASK)
575 
576 static char *nsi_message[30] = {
577  "NSI Link Down", /* NSI_FERR/NSI_NERR bit 0, fatal error */
578  "", /* reserved */
579  "NSI Parity Error", /* bit 2, non-fatal */
580  "", /* reserved */
581  "", /* reserved */
582  "Correctable Error Message", /* bit 5, non-fatal */
583  "Non-Fatal Error Message", /* bit 6, non-fatal */
584  "Fatal Error Message", /* bit 7, fatal */
585  "", /* reserved */
586  "Receiver Error", /* bit 9, non-fatal */
587  "", /* reserved */
588  "Bad TLP", /* bit 11, non-fatal */
589  "Bad DLLP", /* bit 12, non-fatal */
590  "REPLAY_NUM Rollover", /* bit 13, non-fatal */
591  "", /* reserved */
592  "Replay Timer Timeout", /* bit 15, non-fatal */
593  "", /* reserved */
594  "", /* reserved */
595  "", /* reserved */
596  "Data Link Protocol Error", /* bit 19, fatal */
597  "", /* reserved */
598  "Poisoned TLP", /* bit 21, non-fatal */
599  "", /* reserved */
600  "Completion Timeout", /* bit 23, non-fatal */
601  "Completer Abort", /* bit 24, non-fatal */
602  "Unexpected Completion", /* bit 25, non-fatal */
603  "Receiver Overflow", /* bit 26, fatal */
604  "Malformed TLP", /* bit 27, fatal */
605  "", /* reserved */
606  "Unsupported Request" /* bit 29, non-fatal */
607 };
608 
609 static void do_nsi_error(int fatal, u32 errors)
610 {
611  int i;
612 
613  for (i = 0; i < 30; i++) {
614  if (errors & (1 << i))
615  printk(KERN_WARNING "%sError %s\n",
616  fatal_message[fatal], nsi_message[i]);
617  }
618 }
619 
620 static inline void nsi_error(int fatal, u32 errors, int *error_found,
621  int handle_error)
622 {
623  *error_found = 1;
624 
625  if (handle_error)
626  do_nsi_error(fatal, errors);
627 }
628 
629 static char *membuf_message[4] = {
630  "Internal PMWB to DRAM parity",
631  "Internal PMWB to System Bus Parity",
632  "Internal System Bus or IO to PMWB Parity",
633  "Internal DRAM to PMWB Parity"
634 };
635 
636 static void do_membuf_error(u8 errors)
637 {
638  int i;
639 
640  for (i = 0; i < 4; i++) {
641  if (errors & (1 << i))
642  e752x_printk(KERN_WARNING, "Non-Fatal Error %s\n",
643  membuf_message[i]);
644  }
645 }
646 
647 static inline void membuf_error(u8 errors, int *error_found, int handle_error)
648 {
649  *error_found = 1;
650 
651  if (handle_error)
652  do_membuf_error(errors);
653 }
654 
655 static char *sysbus_message[10] = {
656  "Addr or Request Parity",
657  "Data Strobe Glitch",
658  "Addr Strobe Glitch",
659  "Data Parity",
660  "Addr Above TOM",
661  "Non DRAM Lock Error",
662  "MCERR", "BINIT",
663  "Memory Parity",
664  "IO Subsystem Parity"
665 };
666 
667 static void do_sysbus_error(int fatal, u32 errors)
668 {
669  int i;
670 
671  for (i = 0; i < 10; i++) {
672  if (errors & (1 << i))
673  e752x_printk(KERN_WARNING, "%sError System Bus %s\n",
674  fatal_message[fatal], sysbus_message[i]);
675  }
676 }
677 
678 static inline void sysbus_error(int fatal, u32 errors, int *error_found,
679  int handle_error)
680 {
681  *error_found = 1;
682 
683  if (handle_error)
684  do_sysbus_error(fatal, errors);
685 }
686 
687 static void e752x_check_hub_interface(struct e752x_error_info *info,
688  int *error_found, int handle_error)
689 {
690  u8 stat8;
691 
692  //pci_read_config_byte(dev,E752X_HI_FERR,&stat8);
693 
694  stat8 = info->hi_ferr;
695 
696  if (stat8 & 0x7f) { /* Error, so process */
697  stat8 &= 0x7f;
698 
699  if (stat8 & 0x2b)
700  hub_error(1, stat8 & 0x2b, error_found, handle_error);
701 
702  if (stat8 & 0x54)
703  hub_error(0, stat8 & 0x54, error_found, handle_error);
704  }
705  //pci_read_config_byte(dev,E752X_HI_NERR,&stat8);
706 
707  stat8 = info->hi_nerr;
708 
709  if (stat8 & 0x7f) { /* Error, so process */
710  stat8 &= 0x7f;
711 
712  if (stat8 & 0x2b)
713  hub_error(1, stat8 & 0x2b, error_found, handle_error);
714 
715  if (stat8 & 0x54)
716  hub_error(0, stat8 & 0x54, error_found, handle_error);
717  }
718 }
719 
720 static void e752x_check_ns_interface(struct e752x_error_info *info,
721  int *error_found, int handle_error)
722 {
723  u32 stat32;
724 
725  stat32 = info->nsi_ferr;
726  if (stat32 & NSI_ERR_MASK) { /* Error, so process */
727  if (stat32 & NSI_FATAL_MASK) /* check for fatal errors */
728  nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
729  handle_error);
730  if (stat32 & NSI_NON_FATAL_MASK) /* check for non-fatal ones */
731  nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
732  handle_error);
733  }
734  stat32 = info->nsi_nerr;
735  if (stat32 & NSI_ERR_MASK) {
736  if (stat32 & NSI_FATAL_MASK)
737  nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
738  handle_error);
739  if (stat32 & NSI_NON_FATAL_MASK)
740  nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
741  handle_error);
742  }
743 }
744 
745 static void e752x_check_sysbus(struct e752x_error_info *info,
746  int *error_found, int handle_error)
747 {
748  u32 stat32, error32;
749 
750  //pci_read_config_dword(dev,E752X_SYSBUS_FERR,&stat32);
751  stat32 = info->sysbus_ferr + (info->sysbus_nerr << 16);
752 
753  if (stat32 == 0)
754  return; /* no errors */
755 
756  error32 = (stat32 >> 16) & 0x3ff;
757  stat32 = stat32 & 0x3ff;
758 
759  if (stat32 & 0x087)
760  sysbus_error(1, stat32 & 0x087, error_found, handle_error);
761 
762  if (stat32 & 0x378)
763  sysbus_error(0, stat32 & 0x378, error_found, handle_error);
764 
765  if (error32 & 0x087)
766  sysbus_error(1, error32 & 0x087, error_found, handle_error);
767 
768  if (error32 & 0x378)
769  sysbus_error(0, error32 & 0x378, error_found, handle_error);
770 }
771 
772 static void e752x_check_membuf(struct e752x_error_info *info,
773  int *error_found, int handle_error)
774 {
775  u8 stat8;
776 
777  stat8 = info->buf_ferr;
778 
779  if (stat8 & 0x0f) { /* Error, so process */
780  stat8 &= 0x0f;
781  membuf_error(stat8, error_found, handle_error);
782  }
783 
784  stat8 = info->buf_nerr;
785 
786  if (stat8 & 0x0f) { /* Error, so process */
787  stat8 &= 0x0f;
788  membuf_error(stat8, error_found, handle_error);
789  }
790 }
791 
792 static void e752x_check_dram(struct mem_ctl_info *mci,
793  struct e752x_error_info *info, int *error_found,
794  int handle_error)
795 {
796  u16 error_one, error_next;
797 
798  error_one = info->dram_ferr;
799  error_next = info->dram_nerr;
800 
801  /* decode and report errors */
802  if (error_one & 0x0101) /* check first error correctable */
803  process_ce(mci, error_one, info->dram_sec1_add,
804  info->dram_sec1_syndrome, error_found, handle_error);
805 
806  if (error_next & 0x0101) /* check next error correctable */
807  process_ce(mci, error_next, info->dram_sec2_add,
808  info->dram_sec2_syndrome, error_found, handle_error);
809 
810  if (error_one & 0x4040)
811  process_ue_no_info_wr(mci, error_found, handle_error);
812 
813  if (error_next & 0x4040)
814  process_ue_no_info_wr(mci, error_found, handle_error);
815 
816  if (error_one & 0x2020)
817  process_ded_retry(mci, error_one, info->dram_retr_add,
818  error_found, handle_error);
819 
820  if (error_next & 0x2020)
821  process_ded_retry(mci, error_next, info->dram_retr_add,
822  error_found, handle_error);
823 
824  if (error_one & 0x0808)
825  process_threshold_ce(mci, error_one, error_found, handle_error);
826 
827  if (error_next & 0x0808)
828  process_threshold_ce(mci, error_next, error_found,
829  handle_error);
830 
831  if (error_one & 0x0606)
832  process_ue(mci, error_one, info->dram_ded_add,
833  info->dram_scrb_add, error_found, handle_error);
834 
835  if (error_next & 0x0606)
836  process_ue(mci, error_next, info->dram_ded_add,
837  info->dram_scrb_add, error_found, handle_error);
838 }
839 
840 static void e752x_get_error_info(struct mem_ctl_info *mci,
841  struct e752x_error_info *info)
842 {
843  struct pci_dev *dev;
844  struct e752x_pvt *pvt;
845 
846  memset(info, 0, sizeof(*info));
847  pvt = (struct e752x_pvt *)mci->pvt_info;
848  dev = pvt->dev_d0f1;
849  pci_read_config_dword(dev, E752X_FERR_GLOBAL, &info->ferr_global);
850 
851  if (info->ferr_global) {
852  if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
853  pci_read_config_dword(dev, I3100_NSI_FERR,
854  &info->nsi_ferr);
855  info->hi_ferr = 0;
856  } else {
857  pci_read_config_byte(dev, E752X_HI_FERR,
858  &info->hi_ferr);
859  info->nsi_ferr = 0;
860  }
861  pci_read_config_word(dev, E752X_SYSBUS_FERR,
862  &info->sysbus_ferr);
863  pci_read_config_byte(dev, E752X_BUF_FERR, &info->buf_ferr);
864  pci_read_config_word(dev, E752X_DRAM_FERR, &info->dram_ferr);
865  pci_read_config_dword(dev, E752X_DRAM_SEC1_ADD,
866  &info->dram_sec1_add);
867  pci_read_config_word(dev, E752X_DRAM_SEC1_SYNDROME,
868  &info->dram_sec1_syndrome);
869  pci_read_config_dword(dev, E752X_DRAM_DED_ADD,
870  &info->dram_ded_add);
871  pci_read_config_dword(dev, E752X_DRAM_SCRB_ADD,
872  &info->dram_scrb_add);
873  pci_read_config_dword(dev, E752X_DRAM_RETR_ADD,
874  &info->dram_retr_add);
875 
876  /* ignore the reserved bits just in case */
877  if (info->hi_ferr & 0x7f)
878  pci_write_config_byte(dev, E752X_HI_FERR,
879  info->hi_ferr);
880 
881  if (info->nsi_ferr & NSI_ERR_MASK)
882  pci_write_config_dword(dev, I3100_NSI_FERR,
883  info->nsi_ferr);
884 
885  if (info->sysbus_ferr)
886  pci_write_config_word(dev, E752X_SYSBUS_FERR,
887  info->sysbus_ferr);
888 
889  if (info->buf_ferr & 0x0f)
890  pci_write_config_byte(dev, E752X_BUF_FERR,
891  info->buf_ferr);
892 
893  if (info->dram_ferr)
894  pci_write_bits16(pvt->bridge_ck, E752X_DRAM_FERR,
895  info->dram_ferr, info->dram_ferr);
896 
897  pci_write_config_dword(dev, E752X_FERR_GLOBAL,
898  info->ferr_global);
899  }
900 
901  pci_read_config_dword(dev, E752X_NERR_GLOBAL, &info->nerr_global);
902 
903  if (info->nerr_global) {
904  if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
905  pci_read_config_dword(dev, I3100_NSI_NERR,
906  &info->nsi_nerr);
907  info->hi_nerr = 0;
908  } else {
909  pci_read_config_byte(dev, E752X_HI_NERR,
910  &info->hi_nerr);
911  info->nsi_nerr = 0;
912  }
913  pci_read_config_word(dev, E752X_SYSBUS_NERR,
914  &info->sysbus_nerr);
915  pci_read_config_byte(dev, E752X_BUF_NERR, &info->buf_nerr);
916  pci_read_config_word(dev, E752X_DRAM_NERR, &info->dram_nerr);
917  pci_read_config_dword(dev, E752X_DRAM_SEC2_ADD,
918  &info->dram_sec2_add);
919  pci_read_config_word(dev, E752X_DRAM_SEC2_SYNDROME,
920  &info->dram_sec2_syndrome);
921 
922  if (info->hi_nerr & 0x7f)
923  pci_write_config_byte(dev, E752X_HI_NERR,
924  info->hi_nerr);
925 
926  if (info->nsi_nerr & NSI_ERR_MASK)
927  pci_write_config_dword(dev, I3100_NSI_NERR,
928  info->nsi_nerr);
929 
930  if (info->sysbus_nerr)
931  pci_write_config_word(dev, E752X_SYSBUS_NERR,
932  info->sysbus_nerr);
933 
934  if (info->buf_nerr & 0x0f)
935  pci_write_config_byte(dev, E752X_BUF_NERR,
936  info->buf_nerr);
937 
938  if (info->dram_nerr)
939  pci_write_bits16(pvt->bridge_ck, E752X_DRAM_NERR,
940  info->dram_nerr, info->dram_nerr);
941 
942  pci_write_config_dword(dev, E752X_NERR_GLOBAL,
943  info->nerr_global);
944  }
945 }
946 
947 static int e752x_process_error_info(struct mem_ctl_info *mci,
948  struct e752x_error_info *info,
949  int handle_errors)
950 {
951  u32 error32, stat32;
952  int error_found;
953 
954  error_found = 0;
955  error32 = (info->ferr_global >> 18) & 0x3ff;
956  stat32 = (info->ferr_global >> 4) & 0x7ff;
957 
958  if (error32)
959  global_error(1, error32, &error_found, handle_errors);
960 
961  if (stat32)
962  global_error(0, stat32, &error_found, handle_errors);
963 
964  error32 = (info->nerr_global >> 18) & 0x3ff;
965  stat32 = (info->nerr_global >> 4) & 0x7ff;
966 
967  if (error32)
968  global_error(1, error32, &error_found, handle_errors);
969 
970  if (stat32)
971  global_error(0, stat32, &error_found, handle_errors);
972 
973  e752x_check_hub_interface(info, &error_found, handle_errors);
974  e752x_check_ns_interface(info, &error_found, handle_errors);
975  e752x_check_sysbus(info, &error_found, handle_errors);
976  e752x_check_membuf(info, &error_found, handle_errors);
977  e752x_check_dram(mci, info, &error_found, handle_errors);
978  return error_found;
979 }
980 
981 static void e752x_check(struct mem_ctl_info *mci)
982 {
983  struct e752x_error_info info;
984 
985  edac_dbg(3, "\n");
986  e752x_get_error_info(mci, &info);
987  e752x_process_error_info(mci, &info, 1);
988 }
989 
990 /* Program byte/sec bandwidth scrub rate to hardware */
991 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw)
992 {
993  const struct scrubrate *scrubrates;
994  struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
995  struct pci_dev *pdev = pvt->dev_d0f0;
996  int i;
997 
998  if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
999  scrubrates = scrubrates_i3100;
1000  else
1001  scrubrates = scrubrates_e752x;
1002 
1003  /* Translate the desired scrub rate to a e752x/3100 register value.
1004  * Search for the bandwidth that is equal or greater than the
1005  * desired rate and program the cooresponding register value.
1006  */
1007  for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1008  if (scrubrates[i].bandwidth >= new_bw)
1009  break;
1010 
1011  if (scrubrates[i].bandwidth == SDRATE_EOT)
1012  return -1;
1013 
1014  pci_write_config_word(pdev, E752X_MCHSCRB, scrubrates[i].scrubval);
1015 
1016  return scrubrates[i].bandwidth;
1017 }
1018 
1019 /* Convert current scrub rate value into byte/sec bandwidth */
1020 static int get_sdram_scrub_rate(struct mem_ctl_info *mci)
1021 {
1022  const struct scrubrate *scrubrates;
1023  struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
1024  struct pci_dev *pdev = pvt->dev_d0f0;
1025  u16 scrubval;
1026  int i;
1027 
1028  if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
1029  scrubrates = scrubrates_i3100;
1030  else
1031  scrubrates = scrubrates_e752x;
1032 
1033  /* Find the bandwidth matching the memory scrubber configuration */
1034  pci_read_config_word(pdev, E752X_MCHSCRB, &scrubval);
1035  scrubval = scrubval & 0x0f;
1036 
1037  for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1038  if (scrubrates[i].scrubval == scrubval)
1039  break;
1040 
1041  if (scrubrates[i].bandwidth == SDRATE_EOT) {
1043  "Invalid sdram scrub control value: 0x%x\n", scrubval);
1044  return -1;
1045  }
1046  return scrubrates[i].bandwidth;
1047 
1048 }
1049 
1050 /* Return 1 if dual channel mode is active. Else return 0. */
1051 static inline int dual_channel_active(u16 ddrcsr)
1052 {
1053  return (((ddrcsr >> 12) & 3) == 3);
1054 }
1055 
1056 /* Remap csrow index numbers if map_type is "reverse"
1057  */
1058 static inline int remap_csrow_index(struct mem_ctl_info *mci, int index)
1059 {
1060  struct e752x_pvt *pvt = mci->pvt_info;
1061 
1062  if (!pvt->map_type)
1063  return (7 - index);
1064 
1065  return (index);
1066 }
1067 
1068 static void e752x_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev,
1069  u16 ddrcsr)
1070 {
1071  struct csrow_info *csrow;
1072  enum edac_type edac_mode;
1073  unsigned long last_cumul_size;
1074  int index, mem_dev, drc_chan;
1075  int drc_drbg; /* DRB granularity 0=64mb, 1=128mb */
1076  int drc_ddim; /* DRAM Data Integrity Mode 0=none, 2=edac */
1077  u8 value;
1078  u32 dra, drc, cumul_size, i, nr_pages;
1079 
1080  dra = 0;
1081  for (index = 0; index < 4; index++) {
1082  u8 dra_reg;
1083  pci_read_config_byte(pdev, E752X_DRA + index, &dra_reg);
1084  dra |= dra_reg << (index * 8);
1085  }
1086  pci_read_config_dword(pdev, E752X_DRC, &drc);
1087  drc_chan = dual_channel_active(ddrcsr) ? 1 : 0;
1088  drc_drbg = drc_chan + 1; /* 128 in dual mode, 64 in single */
1089  drc_ddim = (drc >> 20) & 0x3;
1090 
1091  /* The dram row boundary (DRB) reg values are boundary address for
1092  * each DRAM row with a granularity of 64 or 128MB (single/dual
1093  * channel operation). DRB regs are cumulative; therefore DRB7 will
1094  * contain the total memory contained in all eight rows.
1095  */
1096  for (last_cumul_size = index = 0; index < mci->nr_csrows; index++) {
1097  /* mem_dev 0=x8, 1=x4 */
1098  mem_dev = (dra >> (index * 4 + 2)) & 0x3;
1099  csrow = mci->csrows[remap_csrow_index(mci, index)];
1100 
1101  mem_dev = (mem_dev == 2);
1102  pci_read_config_byte(pdev, E752X_DRB + index, &value);
1103  /* convert a 128 or 64 MiB DRB to a page size. */
1104  cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
1105  edac_dbg(3, "(%d) cumul_size 0x%x\n", index, cumul_size);
1106  if (cumul_size == last_cumul_size)
1107  continue; /* not populated */
1108 
1109  csrow->first_page = last_cumul_size;
1110  csrow->last_page = cumul_size - 1;
1111  nr_pages = cumul_size - last_cumul_size;
1112  last_cumul_size = cumul_size;
1113 
1114  /*
1115  * if single channel or x8 devices then SECDED
1116  * if dual channel and x4 then S4ECD4ED
1117  */
1118  if (drc_ddim) {
1119  if (drc_chan && mem_dev) {
1120  edac_mode = EDAC_S4ECD4ED;
1121  mci->edac_cap |= EDAC_FLAG_S4ECD4ED;
1122  } else {
1123  edac_mode = EDAC_SECDED;
1124  mci->edac_cap |= EDAC_FLAG_SECDED;
1125  }
1126  } else
1127  edac_mode = EDAC_NONE;
1128  for (i = 0; i < csrow->nr_channels; i++) {
1129  struct dimm_info *dimm = csrow->channels[i]->dimm;
1130 
1131  edac_dbg(3, "Initializing rank at (%i,%i)\n", index, i);
1132  dimm->nr_pages = nr_pages / csrow->nr_channels;
1133  dimm->grain = 1 << 12; /* 4KiB - resolution of CELOG */
1134  dimm->mtype = MEM_RDDR; /* only one type supported */
1135  dimm->dtype = mem_dev ? DEV_X4 : DEV_X8;
1136  dimm->edac_mode = edac_mode;
1137  }
1138  }
1139 }
1140 
1141 static void e752x_init_mem_map_table(struct pci_dev *pdev,
1142  struct e752x_pvt *pvt)
1143 {
1144  int index;
1145  u8 value, last, row;
1146 
1147  last = 0;
1148  row = 0;
1149 
1150  for (index = 0; index < 8; index += 2) {
1151  pci_read_config_byte(pdev, E752X_DRB + index, &value);
1152  /* test if there is a dimm in this slot */
1153  if (value == last) {
1154  /* no dimm in the slot, so flag it as empty */
1155  pvt->map[index] = 0xff;
1156  pvt->map[index + 1] = 0xff;
1157  } else { /* there is a dimm in the slot */
1158  pvt->map[index] = row;
1159  row++;
1160  last = value;
1161  /* test the next value to see if the dimm is double
1162  * sided
1163  */
1164  pci_read_config_byte(pdev, E752X_DRB + index + 1,
1165  &value);
1166 
1167  /* the dimm is single sided, so flag as empty */
1168  /* this is a double sided dimm to save the next row #*/
1169  pvt->map[index + 1] = (value == last) ? 0xff : row;
1170  row++;
1171  last = value;
1172  }
1173  }
1174 }
1175 
1176 /* Return 0 on success or 1 on failure. */
1177 static int e752x_get_devs(struct pci_dev *pdev, int dev_idx,
1178  struct e752x_pvt *pvt)
1179 {
1180  struct pci_dev *dev;
1181 
1183  pvt->dev_info->err_dev, pvt->bridge_ck);
1184 
1185  if (pvt->bridge_ck == NULL)
1186  pvt->bridge_ck = pci_scan_single_device(pdev->bus,
1187  PCI_DEVFN(0, 1));
1188 
1189  if (pvt->bridge_ck == NULL) {
1190  e752x_printk(KERN_ERR, "error reporting device not found:"
1191  "vendor %x device 0x%x (broken BIOS?)\n",
1192  PCI_VENDOR_ID_INTEL, e752x_devs[dev_idx].err_dev);
1193  return 1;
1194  }
1195 
1197  e752x_devs[dev_idx].ctl_dev,
1198  NULL);
1199 
1200  if (dev == NULL)
1201  goto fail;
1202 
1203  pvt->dev_d0f0 = dev;
1204  pvt->dev_d0f1 = pci_dev_get(pvt->bridge_ck);
1205 
1206  return 0;
1207 
1208 fail:
1209  pci_dev_put(pvt->bridge_ck);
1210  return 1;
1211 }
1212 
1213 /* Setup system bus parity mask register.
1214  * Sysbus parity supported on:
1215  * e7320/e7520/e7525 + Xeon
1216  */
1217 static void e752x_init_sysbus_parity_mask(struct e752x_pvt *pvt)
1218 {
1219  char *cpu_id = cpu_data(0).x86_model_id;
1220  struct pci_dev *dev = pvt->dev_d0f1;
1221  int enable = 1;
1222 
1223  /* Allow module parameter override, else see if CPU supports parity */
1224  if (sysbus_parity != -1) {
1225  enable = sysbus_parity;
1226  } else if (cpu_id[0] && !strstr(cpu_id, "Xeon")) {
1227  e752x_printk(KERN_INFO, "System Bus Parity not "
1228  "supported by CPU, disabling\n");
1229  enable = 0;
1230  }
1231 
1232  if (enable)
1233  pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0000);
1234  else
1235  pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0309);
1236 }
1237 
1238 static void e752x_init_error_reporting_regs(struct e752x_pvt *pvt)
1239 {
1240  struct pci_dev *dev;
1241 
1242  dev = pvt->dev_d0f1;
1243  /* Turn off error disable & SMI in case the BIOS turned it on */
1244  if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
1245  pci_write_config_dword(dev, I3100_NSI_EMASK, 0);
1246  pci_write_config_dword(dev, I3100_NSI_SMICMD, 0);
1247  } else {
1248  pci_write_config_byte(dev, E752X_HI_ERRMASK, 0x00);
1249  pci_write_config_byte(dev, E752X_HI_SMICMD, 0x00);
1250  }
1251 
1252  e752x_init_sysbus_parity_mask(pvt);
1253 
1254  pci_write_config_word(dev, E752X_SYSBUS_SMICMD, 0x00);
1255  pci_write_config_byte(dev, E752X_BUF_ERRMASK, 0x00);
1256  pci_write_config_byte(dev, E752X_BUF_SMICMD, 0x00);
1257  pci_write_config_byte(dev, E752X_DRAM_ERRMASK, 0x00);
1258  pci_write_config_byte(dev, E752X_DRAM_SMICMD, 0x00);
1259 }
1260 
1261 static int e752x_probe1(struct pci_dev *pdev, int dev_idx)
1262 {
1263  u16 pci_data;
1264  u8 stat8;
1265  struct mem_ctl_info *mci;
1266  struct edac_mc_layer layers[2];
1267  struct e752x_pvt *pvt;
1268  u16 ddrcsr;
1269  int drc_chan; /* Number of channels 0=1chan,1=2chan */
1270  struct e752x_error_info discard;
1271 
1272  edac_dbg(0, "mci\n");
1273  edac_dbg(0, "Starting Probe1\n");
1274 
1275  /* check to see if device 0 function 1 is enabled; if it isn't, we
1276  * assume the BIOS has reserved it for a reason and is expecting
1277  * exclusive access, we take care not to violate that assumption and
1278  * fail the probe. */
1279  pci_read_config_byte(pdev, E752X_DEVPRES1, &stat8);
1280  if (!force_function_unhide && !(stat8 & (1 << 5))) {
1281  printk(KERN_INFO "Contact your BIOS vendor to see if the "
1282  "E752x error registers can be safely un-hidden\n");
1283  return -ENODEV;
1284  }
1285  stat8 |= (1 << 5);
1286  pci_write_config_byte(pdev, E752X_DEVPRES1, stat8);
1287 
1288  pci_read_config_word(pdev, E752X_DDRCSR, &ddrcsr);
1289  /* FIXME: should check >>12 or 0xf, true for all? */
1290  /* Dual channel = 1, Single channel = 0 */
1291  drc_chan = dual_channel_active(ddrcsr);
1292 
1294  layers[0].size = E752X_NR_CSROWS;
1295  layers[0].is_virt_csrow = true;
1296  layers[1].type = EDAC_MC_LAYER_CHANNEL;
1297  layers[1].size = drc_chan + 1;
1298  layers[1].is_virt_csrow = false;
1299  mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt));
1300  if (mci == NULL)
1301  return -ENOMEM;
1302 
1303  edac_dbg(3, "init mci\n");
1304  mci->mtype_cap = MEM_FLAG_RDDR;
1305  /* 3100 IMCH supports SECDEC only */
1306  mci->edac_ctl_cap = (dev_idx == I3100) ? EDAC_FLAG_SECDED :
1308  /* FIXME - what if different memory types are in different csrows? */
1309  mci->mod_name = EDAC_MOD_STR;
1310  mci->mod_ver = E752X_REVISION;
1311  mci->pdev = &pdev->dev;
1312 
1313  edac_dbg(3, "init pvt\n");
1314  pvt = (struct e752x_pvt *)mci->pvt_info;
1315  pvt->dev_info = &e752x_devs[dev_idx];
1316  pvt->mc_symmetric = ((ddrcsr & 0x10) != 0);
1317 
1318  if (e752x_get_devs(pdev, dev_idx, pvt)) {
1319  edac_mc_free(mci);
1320  return -ENODEV;
1321  }
1322 
1323  edac_dbg(3, "more mci init\n");
1324  mci->ctl_name = pvt->dev_info->ctl_name;
1325  mci->dev_name = pci_name(pdev);
1326  mci->edac_check = e752x_check;
1327  mci->ctl_page_to_phys = ctl_page_to_phys;
1328  mci->set_sdram_scrub_rate = set_sdram_scrub_rate;
1329  mci->get_sdram_scrub_rate = get_sdram_scrub_rate;
1330 
1331  /* set the map type. 1 = normal, 0 = reversed
1332  * Must be set before e752x_init_csrows in case csrow mapping
1333  * is reversed.
1334  */
1335  pci_read_config_byte(pdev, E752X_DRM, &stat8);
1336  pvt->map_type = ((stat8 & 0x0f) > ((stat8 >> 4) & 0x0f));
1337 
1338  e752x_init_csrows(mci, pdev, ddrcsr);
1339  e752x_init_mem_map_table(pdev, pvt);
1340 
1341  if (dev_idx == I3100)
1342  mci->edac_cap = EDAC_FLAG_SECDED; /* the only mode supported */
1343  else
1344  mci->edac_cap |= EDAC_FLAG_NONE;
1345  edac_dbg(3, "tolm, remapbase, remaplimit\n");
1346 
1347  /* load the top of low memory, remap base, and remap limit vars */
1348  pci_read_config_word(pdev, E752X_TOLM, &pci_data);
1349  pvt->tolm = ((u32) pci_data) << 4;
1350  pci_read_config_word(pdev, E752X_REMAPBASE, &pci_data);
1351  pvt->remapbase = ((u32) pci_data) << 14;
1352  pci_read_config_word(pdev, E752X_REMAPLIMIT, &pci_data);
1353  pvt->remaplimit = ((u32) pci_data) << 14;
1355  "tolm = %x, remapbase = %x, remaplimit = %x\n",
1356  pvt->tolm, pvt->remapbase, pvt->remaplimit);
1357 
1358  /* Here we assume that we will never see multiple instances of this
1359  * type of memory controller. The ID is therefore hardcoded to 0.
1360  */
1361  if (edac_mc_add_mc(mci)) {
1362  edac_dbg(3, "failed edac_mc_add_mc()\n");
1363  goto fail;
1364  }
1365 
1366  e752x_init_error_reporting_regs(pvt);
1367  e752x_get_error_info(mci, &discard); /* clear other MCH errors */
1368 
1369  /* allocating generic PCI control info */
1370  e752x_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR);
1371  if (!e752x_pci) {
1373  "%s(): Unable to create PCI control\n", __func__);
1375  "%s(): PCI error report via EDAC not setup\n",
1376  __func__);
1377  }
1378 
1379  /* get this far and it's successful */
1380  edac_dbg(3, "success\n");
1381  return 0;
1382 
1383 fail:
1384  pci_dev_put(pvt->dev_d0f0);
1385  pci_dev_put(pvt->dev_d0f1);
1386  pci_dev_put(pvt->bridge_ck);
1387  edac_mc_free(mci);
1388 
1389  return -ENODEV;
1390 }
1391 
1392 /* returns count (>= 0), or negative on error */
1393 static int __devinit e752x_init_one(struct pci_dev *pdev,
1394  const struct pci_device_id *ent)
1395 {
1396  edac_dbg(0, "\n");
1397 
1398  /* wake up and enable device */
1399  if (pci_enable_device(pdev) < 0)
1400  return -EIO;
1401 
1402  return e752x_probe1(pdev, ent->driver_data);
1403 }
1404 
1405 static void __devexit e752x_remove_one(struct pci_dev *pdev)
1406 {
1407  struct mem_ctl_info *mci;
1408  struct e752x_pvt *pvt;
1409 
1410  edac_dbg(0, "\n");
1411 
1412  if (e752x_pci)
1413  edac_pci_release_generic_ctl(e752x_pci);
1414 
1415  if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL)
1416  return;
1417 
1418  pvt = (struct e752x_pvt *)mci->pvt_info;
1419  pci_dev_put(pvt->dev_d0f0);
1420  pci_dev_put(pvt->dev_d0f1);
1421  pci_dev_put(pvt->bridge_ck);
1422  edac_mc_free(mci);
1423 }
1424 
1425 static DEFINE_PCI_DEVICE_TABLE(e752x_pci_tbl) = {
1426  {
1427  PCI_VEND_DEV(INTEL, 7520_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1428  E7520},
1429  {
1430  PCI_VEND_DEV(INTEL, 7525_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1431  E7525},
1432  {
1433  PCI_VEND_DEV(INTEL, 7320_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1434  E7320},
1435  {
1436  PCI_VEND_DEV(INTEL, 3100_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1437  I3100},
1438  {
1439  0,
1440  } /* 0 terminated list. */
1441 };
1442 
1443 MODULE_DEVICE_TABLE(pci, e752x_pci_tbl);
1444 
1445 static struct pci_driver e752x_driver = {
1446  .name = EDAC_MOD_STR,
1447  .probe = e752x_init_one,
1448  .remove = __devexit_p(e752x_remove_one),
1449  .id_table = e752x_pci_tbl,
1450 };
1451 
1452 static int __init e752x_init(void)
1453 {
1454  int pci_rc;
1455 
1456  edac_dbg(3, "\n");
1457 
1458  /* Ensure that the OPSTATE is set correctly for POLL or NMI */
1459  opstate_init();
1460 
1461  pci_rc = pci_register_driver(&e752x_driver);
1462  return (pci_rc < 0) ? pci_rc : 0;
1463 }
1464 
1465 static void __exit e752x_exit(void)
1466 {
1467  edac_dbg(3, "\n");
1468  pci_unregister_driver(&e752x_driver);
1469 }
1470 
1471 module_init(e752x_init);
1472 module_exit(e752x_exit);
1473 
1474 MODULE_LICENSE("GPL");
1475 MODULE_AUTHOR("Linux Networx (http://lnxi.com) Tom Zimmerman\n");
1476 MODULE_DESCRIPTION("MC support for Intel e752x/3100 memory controllers");
1477 
1478 module_param(force_function_unhide, int, 0444);
1479 MODULE_PARM_DESC(force_function_unhide, "if BIOS sets Dev0:Fun1 up as hidden:"
1480  " 1=force unhide and hope BIOS doesn't fight driver for "
1481  "Dev0:Fun1 access");
1482 
1483 module_param(edac_op_state, int, 0444);
1484 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
1485 
1486 module_param(sysbus_parity, int, 0444);
1487 MODULE_PARM_DESC(sysbus_parity, "0=disable system bus parity checking,"
1488  " 1=enable system bus parity checking, default=auto-detect");
1489 module_param(report_non_memory_errors, int, 0644);
1490 MODULE_PARM_DESC(report_non_memory_errors, "0=disable non-memory error "
1491  "reporting, 1=enable non-memory error reporting");