Root/
1 | /* |
2 | * Intel e752x Memory Controller kernel module |
3 | * (C) 2004 Linux Networx (http://lnxi.com) |
4 | * This file may be distributed under the terms of the |
5 | * GNU General Public License. |
6 | * |
7 | * Implement support for the e7520, E7525, e7320 and i3100 memory controllers. |
8 | * |
9 | * Datasheets: |
10 | * http://www.intel.in/content/www/in/en/chipsets/e7525-memory-controller-hub-datasheet.html |
11 | * ftp://download.intel.com/design/intarch/datashts/31345803.pdf |
12 | * |
13 | * Written by Tom Zimmerman |
14 | * |
15 | * Contributors: |
16 | * Thayne Harbaugh at realmsys.com (?) |
17 | * Wang Zhenyu at intel.com |
18 | * Dave Jiang at mvista.com |
19 | * |
20 | */ |
21 | |
22 | #include <linux/module.h> |
23 | #include <linux/init.h> |
24 | #include <linux/pci.h> |
25 | #include <linux/pci_ids.h> |
26 | #include <linux/edac.h> |
27 | #include "edac_core.h" |
28 | |
29 | #define E752X_REVISION " Ver: 2.0.2" |
30 | #define EDAC_MOD_STR "e752x_edac" |
31 | |
32 | static int report_non_memory_errors; |
33 | static int force_function_unhide; |
34 | static int sysbus_parity = -1; |
35 | |
36 | static struct edac_pci_ctl_info *e752x_pci; |
37 | |
38 | #define e752x_printk(level, fmt, arg...) \ |
39 | edac_printk(level, "e752x", fmt, ##arg) |
40 | |
41 | #define e752x_mc_printk(mci, level, fmt, arg...) \ |
42 | edac_mc_chipset_printk(mci, level, "e752x", fmt, ##arg) |
43 | |
44 | #ifndef PCI_DEVICE_ID_INTEL_7520_0 |
45 | #define PCI_DEVICE_ID_INTEL_7520_0 0x3590 |
46 | #endif /* PCI_DEVICE_ID_INTEL_7520_0 */ |
47 | |
48 | #ifndef PCI_DEVICE_ID_INTEL_7520_1_ERR |
49 | #define PCI_DEVICE_ID_INTEL_7520_1_ERR 0x3591 |
50 | #endif /* PCI_DEVICE_ID_INTEL_7520_1_ERR */ |
51 | |
52 | #ifndef PCI_DEVICE_ID_INTEL_7525_0 |
53 | #define PCI_DEVICE_ID_INTEL_7525_0 0x359E |
54 | #endif /* PCI_DEVICE_ID_INTEL_7525_0 */ |
55 | |
56 | #ifndef PCI_DEVICE_ID_INTEL_7525_1_ERR |
57 | #define PCI_DEVICE_ID_INTEL_7525_1_ERR 0x3593 |
58 | #endif /* PCI_DEVICE_ID_INTEL_7525_1_ERR */ |
59 | |
60 | #ifndef PCI_DEVICE_ID_INTEL_7320_0 |
61 | #define PCI_DEVICE_ID_INTEL_7320_0 0x3592 |
62 | #endif /* PCI_DEVICE_ID_INTEL_7320_0 */ |
63 | |
64 | #ifndef PCI_DEVICE_ID_INTEL_7320_1_ERR |
65 | #define PCI_DEVICE_ID_INTEL_7320_1_ERR 0x3593 |
66 | #endif /* PCI_DEVICE_ID_INTEL_7320_1_ERR */ |
67 | |
68 | #ifndef PCI_DEVICE_ID_INTEL_3100_0 |
69 | #define PCI_DEVICE_ID_INTEL_3100_0 0x35B0 |
70 | #endif /* PCI_DEVICE_ID_INTEL_3100_0 */ |
71 | |
72 | #ifndef PCI_DEVICE_ID_INTEL_3100_1_ERR |
73 | #define PCI_DEVICE_ID_INTEL_3100_1_ERR 0x35B1 |
74 | #endif /* PCI_DEVICE_ID_INTEL_3100_1_ERR */ |
75 | |
76 | #define E752X_NR_CSROWS 8 /* number of csrows */ |
77 | |
78 | /* E752X register addresses - device 0 function 0 */ |
79 | #define E752X_MCHSCRB 0x52 /* Memory Scrub register (16b) */ |
80 | /* |
81 | * 6:5 Scrub Completion Count |
82 | * 3:2 Scrub Rate (i3100 only) |
83 | * 01=fast 10=normal |
84 | * 1:0 Scrub Mode enable |
85 | * 00=off 10=on |
86 | */ |
87 | #define E752X_DRB 0x60 /* DRAM row boundary register (8b) */ |
88 | #define E752X_DRA 0x70 /* DRAM row attribute register (8b) */ |
89 | /* |
90 | * 31:30 Device width row 7 |
91 | * 01=x8 10=x4 11=x8 DDR2 |
92 | * 27:26 Device width row 6 |
93 | * 23:22 Device width row 5 |
94 | * 19:20 Device width row 4 |
95 | * 15:14 Device width row 3 |
96 | * 11:10 Device width row 2 |
97 | * 7:6 Device width row 1 |
98 | * 3:2 Device width row 0 |
99 | */ |
100 | #define E752X_DRC 0x7C /* DRAM controller mode reg (32b) */ |
101 | /* FIXME:IS THIS RIGHT? */ |
102 | /* |
103 | * 22 Number channels 0=1,1=2 |
104 | * 19:18 DRB Granularity 32/64MB |
105 | */ |
106 | #define E752X_DRM 0x80 /* Dimm mapping register */ |
107 | #define E752X_DDRCSR 0x9A /* DDR control and status reg (16b) */ |
108 | /* |
109 | * 14:12 1 single A, 2 single B, 3 dual |
110 | */ |
111 | #define E752X_TOLM 0xC4 /* DRAM top of low memory reg (16b) */ |
112 | #define E752X_REMAPBASE 0xC6 /* DRAM remap base address reg (16b) */ |
113 | #define E752X_REMAPLIMIT 0xC8 /* DRAM remap limit address reg (16b) */ |
114 | #define E752X_REMAPOFFSET 0xCA /* DRAM remap limit offset reg (16b) */ |
115 | |
116 | /* E752X register addresses - device 0 function 1 */ |
117 | #define E752X_FERR_GLOBAL 0x40 /* Global first error register (32b) */ |
118 | #define E752X_NERR_GLOBAL 0x44 /* Global next error register (32b) */ |
119 | #define E752X_HI_FERR 0x50 /* Hub interface first error reg (8b) */ |
120 | #define E752X_HI_NERR 0x52 /* Hub interface next error reg (8b) */ |
121 | #define E752X_HI_ERRMASK 0x54 /* Hub interface error mask reg (8b) */ |
122 | #define E752X_HI_SMICMD 0x5A /* Hub interface SMI command reg (8b) */ |
123 | #define E752X_SYSBUS_FERR 0x60 /* System buss first error reg (16b) */ |
124 | #define E752X_SYSBUS_NERR 0x62 /* System buss next error reg (16b) */ |
125 | #define E752X_SYSBUS_ERRMASK 0x64 /* System buss error mask reg (16b) */ |
126 | #define E752X_SYSBUS_SMICMD 0x6A /* System buss SMI command reg (16b) */ |
127 | #define E752X_BUF_FERR 0x70 /* Memory buffer first error reg (8b) */ |
128 | #define E752X_BUF_NERR 0x72 /* Memory buffer next error reg (8b) */ |
129 | #define E752X_BUF_ERRMASK 0x74 /* Memory buffer error mask reg (8b) */ |
130 | #define E752X_BUF_SMICMD 0x7A /* Memory buffer SMI cmd reg (8b) */ |
131 | #define E752X_DRAM_FERR 0x80 /* DRAM first error register (16b) */ |
132 | #define E752X_DRAM_NERR 0x82 /* DRAM next error register (16b) */ |
133 | #define E752X_DRAM_ERRMASK 0x84 /* DRAM error mask register (8b) */ |
134 | #define E752X_DRAM_SMICMD 0x8A /* DRAM SMI command register (8b) */ |
135 | #define E752X_DRAM_RETR_ADD 0xAC /* DRAM Retry address register (32b) */ |
136 | #define E752X_DRAM_SEC1_ADD 0xA0 /* DRAM first correctable memory */ |
137 | /* error address register (32b) */ |
138 | /* |
139 | * 31 Reserved |
140 | * 30:2 CE address (64 byte block 34:6 |
141 | * 1 Reserved |
142 | * 0 HiLoCS |
143 | */ |
144 | #define E752X_DRAM_SEC2_ADD 0xC8 /* DRAM first correctable memory */ |
145 | /* error address register (32b) */ |
146 | /* |
147 | * 31 Reserved |
148 | * 30:2 CE address (64 byte block 34:6) |
149 | * 1 Reserved |
150 | * 0 HiLoCS |
151 | */ |
152 | #define E752X_DRAM_DED_ADD 0xA4 /* DRAM first uncorrectable memory */ |
153 | /* error address register (32b) */ |
154 | /* |
155 | * 31 Reserved |
156 | * 30:2 CE address (64 byte block 34:6) |
157 | * 1 Reserved |
158 | * 0 HiLoCS |
159 | */ |
160 | #define E752X_DRAM_SCRB_ADD 0xA8 /* DRAM 1st uncorrectable scrub mem */ |
161 | /* error address register (32b) */ |
162 | /* |
163 | * 31 Reserved |
164 | * 30:2 CE address (64 byte block 34:6 |
165 | * 1 Reserved |
166 | * 0 HiLoCS |
167 | */ |
168 | #define E752X_DRAM_SEC1_SYNDROME 0xC4 /* DRAM first correctable memory */ |
169 | /* error syndrome register (16b) */ |
170 | #define E752X_DRAM_SEC2_SYNDROME 0xC6 /* DRAM second correctable memory */ |
171 | /* error syndrome register (16b) */ |
172 | #define E752X_DEVPRES1 0xF4 /* Device Present 1 register (8b) */ |
173 | |
174 | /* 3100 IMCH specific register addresses - device 0 function 1 */ |
175 | #define I3100_NSI_FERR 0x48 /* NSI first error reg (32b) */ |
176 | #define I3100_NSI_NERR 0x4C /* NSI next error reg (32b) */ |
177 | #define I3100_NSI_SMICMD 0x54 /* NSI SMI command register (32b) */ |
178 | #define I3100_NSI_EMASK 0x90 /* NSI error mask register (32b) */ |
179 | |
180 | /* ICH5R register addresses - device 30 function 0 */ |
181 | #define ICH5R_PCI_STAT 0x06 /* PCI status register (16b) */ |
182 | #define ICH5R_PCI_2ND_STAT 0x1E /* PCI status secondary reg (16b) */ |
183 | #define ICH5R_PCI_BRIDGE_CTL 0x3E /* PCI bridge control register (16b) */ |
184 | |
185 | enum e752x_chips { |
186 | E7520 = 0, |
187 | E7525 = 1, |
188 | E7320 = 2, |
189 | I3100 = 3 |
190 | }; |
191 | |
192 | /* |
193 | * Those chips Support single-rank and dual-rank memories only. |
194 | * |
195 | * On e752x chips, the odd rows are present only on dual-rank memories. |
196 | * Dividing the rank by two will provide the dimm# |
197 | * |
198 | * i3100 MC has a different mapping: it supports only 4 ranks. |
199 | * |
200 | * The mapping is (from 1 to n): |
201 | * slot single-ranked double-ranked |
202 | * dimm #1 -> rank #4 NA |
203 | * dimm #2 -> rank #3 NA |
204 | * dimm #3 -> rank #2 Ranks 2 and 3 |
205 | * dimm #4 -> rank $1 Ranks 1 and 4 |
206 | * |
207 | * FIXME: The current mapping for i3100 considers that it supports up to 8 |
208 | * ranks/chanel, but datasheet says that the MC supports only 4 ranks. |
209 | */ |
210 | |
211 | struct e752x_pvt { |
212 | struct pci_dev *bridge_ck; |
213 | struct pci_dev *dev_d0f0; |
214 | struct pci_dev *dev_d0f1; |
215 | u32 tolm; |
216 | u32 remapbase; |
217 | u32 remaplimit; |
218 | int mc_symmetric; |
219 | u8 map[8]; |
220 | int map_type; |
221 | const struct e752x_dev_info *dev_info; |
222 | }; |
223 | |
224 | struct e752x_dev_info { |
225 | u16 err_dev; |
226 | u16 ctl_dev; |
227 | const char *ctl_name; |
228 | }; |
229 | |
230 | struct e752x_error_info { |
231 | u32 ferr_global; |
232 | u32 nerr_global; |
233 | u32 nsi_ferr; /* 3100 only */ |
234 | u32 nsi_nerr; /* 3100 only */ |
235 | u8 hi_ferr; /* all but 3100 */ |
236 | u8 hi_nerr; /* all but 3100 */ |
237 | u16 sysbus_ferr; |
238 | u16 sysbus_nerr; |
239 | u8 buf_ferr; |
240 | u8 buf_nerr; |
241 | u16 dram_ferr; |
242 | u16 dram_nerr; |
243 | u32 dram_sec1_add; |
244 | u32 dram_sec2_add; |
245 | u16 dram_sec1_syndrome; |
246 | u16 dram_sec2_syndrome; |
247 | u32 dram_ded_add; |
248 | u32 dram_scrb_add; |
249 | u32 dram_retr_add; |
250 | }; |
251 | |
252 | static const struct e752x_dev_info e752x_devs[] = { |
253 | [E7520] = { |
254 | .err_dev = PCI_DEVICE_ID_INTEL_7520_1_ERR, |
255 | .ctl_dev = PCI_DEVICE_ID_INTEL_7520_0, |
256 | .ctl_name = "E7520"}, |
257 | [E7525] = { |
258 | .err_dev = PCI_DEVICE_ID_INTEL_7525_1_ERR, |
259 | .ctl_dev = PCI_DEVICE_ID_INTEL_7525_0, |
260 | .ctl_name = "E7525"}, |
261 | [E7320] = { |
262 | .err_dev = PCI_DEVICE_ID_INTEL_7320_1_ERR, |
263 | .ctl_dev = PCI_DEVICE_ID_INTEL_7320_0, |
264 | .ctl_name = "E7320"}, |
265 | [I3100] = { |
266 | .err_dev = PCI_DEVICE_ID_INTEL_3100_1_ERR, |
267 | .ctl_dev = PCI_DEVICE_ID_INTEL_3100_0, |
268 | .ctl_name = "3100"}, |
269 | }; |
270 | |
271 | /* Valid scrub rates for the e752x/3100 hardware memory scrubber. We |
272 | * map the scrubbing bandwidth to a hardware register value. The 'set' |
273 | * operation finds the 'matching or higher value'. Note that scrubbing |
274 | * on the e752x can only be enabled/disabled. The 3100 supports |
275 | * a normal and fast mode. |
276 | */ |
277 | |
278 | #define SDRATE_EOT 0xFFFFFFFF |
279 | |
280 | struct scrubrate { |
281 | u32 bandwidth; /* bandwidth consumed by scrubbing in bytes/sec */ |
282 | u16 scrubval; /* register value for scrub rate */ |
283 | }; |
284 | |
285 | /* Rate below assumes same performance as i3100 using PC3200 DDR2 in |
286 | * normal mode. e752x bridges don't support choosing normal or fast mode, |
287 | * so the scrubbing bandwidth value isn't all that important - scrubbing is |
288 | * either on or off. |
289 | */ |
290 | static const struct scrubrate scrubrates_e752x[] = { |
291 | {0, 0x00}, /* Scrubbing Off */ |
292 | {500000, 0x02}, /* Scrubbing On */ |
293 | {SDRATE_EOT, 0x00} /* End of Table */ |
294 | }; |
295 | |
296 | /* Fast mode: 2 GByte PC3200 DDR2 scrubbed in 33s = 63161283 bytes/s |
297 | * Normal mode: 125 (32000 / 256) times slower than fast mode. |
298 | */ |
299 | static const struct scrubrate scrubrates_i3100[] = { |
300 | {0, 0x00}, /* Scrubbing Off */ |
301 | {500000, 0x0a}, /* Normal mode - 32k clocks */ |
302 | {62500000, 0x06}, /* Fast mode - 256 clocks */ |
303 | {SDRATE_EOT, 0x00} /* End of Table */ |
304 | }; |
305 | |
306 | static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci, |
307 | unsigned long page) |
308 | { |
309 | u32 remap; |
310 | struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info; |
311 | |
312 | edac_dbg(3, "\n"); |
313 | |
314 | if (page < pvt->tolm) |
315 | return page; |
316 | |
317 | if ((page >= 0x100000) && (page < pvt->remapbase)) |
318 | return page; |
319 | |
320 | remap = (page - pvt->tolm) + pvt->remapbase; |
321 | |
322 | if (remap < pvt->remaplimit) |
323 | return remap; |
324 | |
325 | e752x_printk(KERN_ERR, "Invalid page %lx - out of range\n", page); |
326 | return pvt->tolm - 1; |
327 | } |
328 | |
329 | static void do_process_ce(struct mem_ctl_info *mci, u16 error_one, |
330 | u32 sec1_add, u16 sec1_syndrome) |
331 | { |
332 | u32 page; |
333 | int row; |
334 | int channel; |
335 | int i; |
336 | struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info; |
337 | |
338 | edac_dbg(3, "\n"); |
339 | |
340 | /* convert the addr to 4k page */ |
341 | page = sec1_add >> (PAGE_SHIFT - 4); |
342 | |
343 | /* FIXME - check for -1 */ |
344 | if (pvt->mc_symmetric) { |
345 | /* chip select are bits 14 & 13 */ |
346 | row = ((page >> 1) & 3); |
347 | e752x_printk(KERN_WARNING, |
348 | "Test row %d Table %d %d %d %d %d %d %d %d\n", row, |
349 | pvt->map[0], pvt->map[1], pvt->map[2], pvt->map[3], |
350 | pvt->map[4], pvt->map[5], pvt->map[6], |
351 | pvt->map[7]); |
352 | |
353 | /* test for channel remapping */ |
354 | for (i = 0; i < 8; i++) { |
355 | if (pvt->map[i] == row) |
356 | break; |
357 | } |
358 | |
359 | e752x_printk(KERN_WARNING, "Test computed row %d\n", i); |
360 | |
361 | if (i < 8) |
362 | row = i; |
363 | else |
364 | e752x_mc_printk(mci, KERN_WARNING, |
365 | "row %d not found in remap table\n", |
366 | row); |
367 | } else |
368 | row = edac_mc_find_csrow_by_page(mci, page); |
369 | |
370 | /* 0 = channel A, 1 = channel B */ |
371 | channel = !(error_one & 1); |
372 | |
373 | /* e752x mc reads 34:6 of the DRAM linear address */ |
374 | edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1, |
375 | page, offset_in_page(sec1_add << 4), sec1_syndrome, |
376 | row, channel, -1, |
377 | "e752x CE", ""); |
378 | } |
379 | |
380 | static inline void process_ce(struct mem_ctl_info *mci, u16 error_one, |
381 | u32 sec1_add, u16 sec1_syndrome, int *error_found, |
382 | int handle_error) |
383 | { |
384 | *error_found = 1; |
385 | |
386 | if (handle_error) |
387 | do_process_ce(mci, error_one, sec1_add, sec1_syndrome); |
388 | } |
389 | |
390 | static void do_process_ue(struct mem_ctl_info *mci, u16 error_one, |
391 | u32 ded_add, u32 scrb_add) |
392 | { |
393 | u32 error_2b, block_page; |
394 | int row; |
395 | struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info; |
396 | |
397 | edac_dbg(3, "\n"); |
398 | |
399 | if (error_one & 0x0202) { |
400 | error_2b = ded_add; |
401 | |
402 | /* convert to 4k address */ |
403 | block_page = error_2b >> (PAGE_SHIFT - 4); |
404 | |
405 | row = pvt->mc_symmetric ? |
406 | /* chip select are bits 14 & 13 */ |
407 | ((block_page >> 1) & 3) : |
408 | edac_mc_find_csrow_by_page(mci, block_page); |
409 | |
410 | /* e752x mc reads 34:6 of the DRAM linear address */ |
411 | edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, |
412 | block_page, |
413 | offset_in_page(error_2b << 4), 0, |
414 | row, -1, -1, |
415 | "e752x UE from Read", ""); |
416 | |
417 | } |
418 | if (error_one & 0x0404) { |
419 | error_2b = scrb_add; |
420 | |
421 | /* convert to 4k address */ |
422 | block_page = error_2b >> (PAGE_SHIFT - 4); |
423 | |
424 | row = pvt->mc_symmetric ? |
425 | /* chip select are bits 14 & 13 */ |
426 | ((block_page >> 1) & 3) : |
427 | edac_mc_find_csrow_by_page(mci, block_page); |
428 | |
429 | /* e752x mc reads 34:6 of the DRAM linear address */ |
430 | edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, |
431 | block_page, |
432 | offset_in_page(error_2b << 4), 0, |
433 | row, -1, -1, |
434 | "e752x UE from Scruber", ""); |
435 | } |
436 | } |
437 | |
438 | static inline void process_ue(struct mem_ctl_info *mci, u16 error_one, |
439 | u32 ded_add, u32 scrb_add, int *error_found, |
440 | int handle_error) |
441 | { |
442 | *error_found = 1; |
443 | |
444 | if (handle_error) |
445 | do_process_ue(mci, error_one, ded_add, scrb_add); |
446 | } |
447 | |
448 | static inline void process_ue_no_info_wr(struct mem_ctl_info *mci, |
449 | int *error_found, int handle_error) |
450 | { |
451 | *error_found = 1; |
452 | |
453 | if (!handle_error) |
454 | return; |
455 | |
456 | edac_dbg(3, "\n"); |
457 | edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, 0, 0, 0, |
458 | -1, -1, -1, |
459 | "e752x UE log memory write", ""); |
460 | } |
461 | |
462 | static void do_process_ded_retry(struct mem_ctl_info *mci, u16 error, |
463 | u32 retry_add) |
464 | { |
465 | u32 error_1b, page; |
466 | int row; |
467 | struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info; |
468 | |
469 | error_1b = retry_add; |
470 | page = error_1b >> (PAGE_SHIFT - 4); /* convert the addr to 4k page */ |
471 | |
472 | /* chip select are bits 14 & 13 */ |
473 | row = pvt->mc_symmetric ? ((page >> 1) & 3) : |
474 | edac_mc_find_csrow_by_page(mci, page); |
475 | |
476 | e752x_mc_printk(mci, KERN_WARNING, |
477 | "CE page 0x%lx, row %d : Memory read retry\n", |
478 | (long unsigned int)page, row); |
479 | } |
480 | |
481 | static inline void process_ded_retry(struct mem_ctl_info *mci, u16 error, |
482 | u32 retry_add, int *error_found, |
483 | int handle_error) |
484 | { |
485 | *error_found = 1; |
486 | |
487 | if (handle_error) |
488 | do_process_ded_retry(mci, error, retry_add); |
489 | } |
490 | |
491 | static inline void process_threshold_ce(struct mem_ctl_info *mci, u16 error, |
492 | int *error_found, int handle_error) |
493 | { |
494 | *error_found = 1; |
495 | |
496 | if (handle_error) |
497 | e752x_mc_printk(mci, KERN_WARNING, "Memory threshold CE\n"); |
498 | } |
499 | |
500 | static char *global_message[11] = { |
501 | "PCI Express C1", |
502 | "PCI Express C", |
503 | "PCI Express B1", |
504 | "PCI Express B", |
505 | "PCI Express A1", |
506 | "PCI Express A", |
507 | "DMA Controller", |
508 | "HUB or NS Interface", |
509 | "System Bus", |
510 | "DRAM Controller", /* 9th entry */ |
511 | "Internal Buffer" |
512 | }; |
513 | |
514 | #define DRAM_ENTRY 9 |
515 | |
516 | static char *fatal_message[2] = { "Non-Fatal ", "Fatal " }; |
517 | |
518 | static void do_global_error(int fatal, u32 errors) |
519 | { |
520 | int i; |
521 | |
522 | for (i = 0; i < 11; i++) { |
523 | if (errors & (1 << i)) { |
524 | /* If the error is from DRAM Controller OR |
525 | * we are to report ALL errors, then |
526 | * report the error |
527 | */ |
528 | if ((i == DRAM_ENTRY) || report_non_memory_errors) |
529 | e752x_printk(KERN_WARNING, "%sError %s\n", |
530 | fatal_message[fatal], |
531 | global_message[i]); |
532 | } |
533 | } |
534 | } |
535 | |
536 | static inline void global_error(int fatal, u32 errors, int *error_found, |
537 | int handle_error) |
538 | { |
539 | *error_found = 1; |
540 | |
541 | if (handle_error) |
542 | do_global_error(fatal, errors); |
543 | } |
544 | |
545 | static char *hub_message[7] = { |
546 | "HI Address or Command Parity", "HI Illegal Access", |
547 | "HI Internal Parity", "Out of Range Access", |
548 | "HI Data Parity", "Enhanced Config Access", |
549 | "Hub Interface Target Abort" |
550 | }; |
551 | |
552 | static void do_hub_error(int fatal, u8 errors) |
553 | { |
554 | int i; |
555 | |
556 | for (i = 0; i < 7; i++) { |
557 | if (errors & (1 << i)) |
558 | e752x_printk(KERN_WARNING, "%sError %s\n", |
559 | fatal_message[fatal], hub_message[i]); |
560 | } |
561 | } |
562 | |
563 | static inline void hub_error(int fatal, u8 errors, int *error_found, |
564 | int handle_error) |
565 | { |
566 | *error_found = 1; |
567 | |
568 | if (handle_error) |
569 | do_hub_error(fatal, errors); |
570 | } |
571 | |
572 | #define NSI_FATAL_MASK 0x0c080081 |
573 | #define NSI_NON_FATAL_MASK 0x23a0ba64 |
574 | #define NSI_ERR_MASK (NSI_FATAL_MASK | NSI_NON_FATAL_MASK) |
575 | |
576 | static char *nsi_message[30] = { |
577 | "NSI Link Down", /* NSI_FERR/NSI_NERR bit 0, fatal error */ |
578 | "", /* reserved */ |
579 | "NSI Parity Error", /* bit 2, non-fatal */ |
580 | "", /* reserved */ |
581 | "", /* reserved */ |
582 | "Correctable Error Message", /* bit 5, non-fatal */ |
583 | "Non-Fatal Error Message", /* bit 6, non-fatal */ |
584 | "Fatal Error Message", /* bit 7, fatal */ |
585 | "", /* reserved */ |
586 | "Receiver Error", /* bit 9, non-fatal */ |
587 | "", /* reserved */ |
588 | "Bad TLP", /* bit 11, non-fatal */ |
589 | "Bad DLLP", /* bit 12, non-fatal */ |
590 | "REPLAY_NUM Rollover", /* bit 13, non-fatal */ |
591 | "", /* reserved */ |
592 | "Replay Timer Timeout", /* bit 15, non-fatal */ |
593 | "", /* reserved */ |
594 | "", /* reserved */ |
595 | "", /* reserved */ |
596 | "Data Link Protocol Error", /* bit 19, fatal */ |
597 | "", /* reserved */ |
598 | "Poisoned TLP", /* bit 21, non-fatal */ |
599 | "", /* reserved */ |
600 | "Completion Timeout", /* bit 23, non-fatal */ |
601 | "Completer Abort", /* bit 24, non-fatal */ |
602 | "Unexpected Completion", /* bit 25, non-fatal */ |
603 | "Receiver Overflow", /* bit 26, fatal */ |
604 | "Malformed TLP", /* bit 27, fatal */ |
605 | "", /* reserved */ |
606 | "Unsupported Request" /* bit 29, non-fatal */ |
607 | }; |
608 | |
609 | static void do_nsi_error(int fatal, u32 errors) |
610 | { |
611 | int i; |
612 | |
613 | for (i = 0; i < 30; i++) { |
614 | if (errors & (1 << i)) |
615 | printk(KERN_WARNING "%sError %s\n", |
616 | fatal_message[fatal], nsi_message[i]); |
617 | } |
618 | } |
619 | |
620 | static inline void nsi_error(int fatal, u32 errors, int *error_found, |
621 | int handle_error) |
622 | { |
623 | *error_found = 1; |
624 | |
625 | if (handle_error) |
626 | do_nsi_error(fatal, errors); |
627 | } |
628 | |
629 | static char *membuf_message[4] = { |
630 | "Internal PMWB to DRAM parity", |
631 | "Internal PMWB to System Bus Parity", |
632 | "Internal System Bus or IO to PMWB Parity", |
633 | "Internal DRAM to PMWB Parity" |
634 | }; |
635 | |
636 | static void do_membuf_error(u8 errors) |
637 | { |
638 | int i; |
639 | |
640 | for (i = 0; i < 4; i++) { |
641 | if (errors & (1 << i)) |
642 | e752x_printk(KERN_WARNING, "Non-Fatal Error %s\n", |
643 | membuf_message[i]); |
644 | } |
645 | } |
646 | |
647 | static inline void membuf_error(u8 errors, int *error_found, int handle_error) |
648 | { |
649 | *error_found = 1; |
650 | |
651 | if (handle_error) |
652 | do_membuf_error(errors); |
653 | } |
654 | |
655 | static char *sysbus_message[10] = { |
656 | "Addr or Request Parity", |
657 | "Data Strobe Glitch", |
658 | "Addr Strobe Glitch", |
659 | "Data Parity", |
660 | "Addr Above TOM", |
661 | "Non DRAM Lock Error", |
662 | "MCERR", "BINIT", |
663 | "Memory Parity", |
664 | "IO Subsystem Parity" |
665 | }; |
666 | |
667 | static void do_sysbus_error(int fatal, u32 errors) |
668 | { |
669 | int i; |
670 | |
671 | for (i = 0; i < 10; i++) { |
672 | if (errors & (1 << i)) |
673 | e752x_printk(KERN_WARNING, "%sError System Bus %s\n", |
674 | fatal_message[fatal], sysbus_message[i]); |
675 | } |
676 | } |
677 | |
678 | static inline void sysbus_error(int fatal, u32 errors, int *error_found, |
679 | int handle_error) |
680 | { |
681 | *error_found = 1; |
682 | |
683 | if (handle_error) |
684 | do_sysbus_error(fatal, errors); |
685 | } |
686 | |
687 | static void e752x_check_hub_interface(struct e752x_error_info *info, |
688 | int *error_found, int handle_error) |
689 | { |
690 | u8 stat8; |
691 | |
692 | //pci_read_config_byte(dev,E752X_HI_FERR,&stat8); |
693 | |
694 | stat8 = info->hi_ferr; |
695 | |
696 | if (stat8 & 0x7f) { /* Error, so process */ |
697 | stat8 &= 0x7f; |
698 | |
699 | if (stat8 & 0x2b) |
700 | hub_error(1, stat8 & 0x2b, error_found, handle_error); |
701 | |
702 | if (stat8 & 0x54) |
703 | hub_error(0, stat8 & 0x54, error_found, handle_error); |
704 | } |
705 | //pci_read_config_byte(dev,E752X_HI_NERR,&stat8); |
706 | |
707 | stat8 = info->hi_nerr; |
708 | |
709 | if (stat8 & 0x7f) { /* Error, so process */ |
710 | stat8 &= 0x7f; |
711 | |
712 | if (stat8 & 0x2b) |
713 | hub_error(1, stat8 & 0x2b, error_found, handle_error); |
714 | |
715 | if (stat8 & 0x54) |
716 | hub_error(0, stat8 & 0x54, error_found, handle_error); |
717 | } |
718 | } |
719 | |
720 | static void e752x_check_ns_interface(struct e752x_error_info *info, |
721 | int *error_found, int handle_error) |
722 | { |
723 | u32 stat32; |
724 | |
725 | stat32 = info->nsi_ferr; |
726 | if (stat32 & NSI_ERR_MASK) { /* Error, so process */ |
727 | if (stat32 & NSI_FATAL_MASK) /* check for fatal errors */ |
728 | nsi_error(1, stat32 & NSI_FATAL_MASK, error_found, |
729 | handle_error); |
730 | if (stat32 & NSI_NON_FATAL_MASK) /* check for non-fatal ones */ |
731 | nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found, |
732 | handle_error); |
733 | } |
734 | stat32 = info->nsi_nerr; |
735 | if (stat32 & NSI_ERR_MASK) { |
736 | if (stat32 & NSI_FATAL_MASK) |
737 | nsi_error(1, stat32 & NSI_FATAL_MASK, error_found, |
738 | handle_error); |
739 | if (stat32 & NSI_NON_FATAL_MASK) |
740 | nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found, |
741 | handle_error); |
742 | } |
743 | } |
744 | |
745 | static void e752x_check_sysbus(struct e752x_error_info *info, |
746 | int *error_found, int handle_error) |
747 | { |
748 | u32 stat32, error32; |
749 | |
750 | //pci_read_config_dword(dev,E752X_SYSBUS_FERR,&stat32); |
751 | stat32 = info->sysbus_ferr + (info->sysbus_nerr << 16); |
752 | |
753 | if (stat32 == 0) |
754 | return; /* no errors */ |
755 | |
756 | error32 = (stat32 >> 16) & 0x3ff; |
757 | stat32 = stat32 & 0x3ff; |
758 | |
759 | if (stat32 & 0x087) |
760 | sysbus_error(1, stat32 & 0x087, error_found, handle_error); |
761 | |
762 | if (stat32 & 0x378) |
763 | sysbus_error(0, stat32 & 0x378, error_found, handle_error); |
764 | |
765 | if (error32 & 0x087) |
766 | sysbus_error(1, error32 & 0x087, error_found, handle_error); |
767 | |
768 | if (error32 & 0x378) |
769 | sysbus_error(0, error32 & 0x378, error_found, handle_error); |
770 | } |
771 | |
772 | static void e752x_check_membuf(struct e752x_error_info *info, |
773 | int *error_found, int handle_error) |
774 | { |
775 | u8 stat8; |
776 | |
777 | stat8 = info->buf_ferr; |
778 | |
779 | if (stat8 & 0x0f) { /* Error, so process */ |
780 | stat8 &= 0x0f; |
781 | membuf_error(stat8, error_found, handle_error); |
782 | } |
783 | |
784 | stat8 = info->buf_nerr; |
785 | |
786 | if (stat8 & 0x0f) { /* Error, so process */ |
787 | stat8 &= 0x0f; |
788 | membuf_error(stat8, error_found, handle_error); |
789 | } |
790 | } |
791 | |
792 | static void e752x_check_dram(struct mem_ctl_info *mci, |
793 | struct e752x_error_info *info, int *error_found, |
794 | int handle_error) |
795 | { |
796 | u16 error_one, error_next; |
797 | |
798 | error_one = info->dram_ferr; |
799 | error_next = info->dram_nerr; |
800 | |
801 | /* decode and report errors */ |
802 | if (error_one & 0x0101) /* check first error correctable */ |
803 | process_ce(mci, error_one, info->dram_sec1_add, |
804 | info->dram_sec1_syndrome, error_found, handle_error); |
805 | |
806 | if (error_next & 0x0101) /* check next error correctable */ |
807 | process_ce(mci, error_next, info->dram_sec2_add, |
808 | info->dram_sec2_syndrome, error_found, handle_error); |
809 | |
810 | if (error_one & 0x4040) |
811 | process_ue_no_info_wr(mci, error_found, handle_error); |
812 | |
813 | if (error_next & 0x4040) |
814 | process_ue_no_info_wr(mci, error_found, handle_error); |
815 | |
816 | if (error_one & 0x2020) |
817 | process_ded_retry(mci, error_one, info->dram_retr_add, |
818 | error_found, handle_error); |
819 | |
820 | if (error_next & 0x2020) |
821 | process_ded_retry(mci, error_next, info->dram_retr_add, |
822 | error_found, handle_error); |
823 | |
824 | if (error_one & 0x0808) |
825 | process_threshold_ce(mci, error_one, error_found, handle_error); |
826 | |
827 | if (error_next & 0x0808) |
828 | process_threshold_ce(mci, error_next, error_found, |
829 | handle_error); |
830 | |
831 | if (error_one & 0x0606) |
832 | process_ue(mci, error_one, info->dram_ded_add, |
833 | info->dram_scrb_add, error_found, handle_error); |
834 | |
835 | if (error_next & 0x0606) |
836 | process_ue(mci, error_next, info->dram_ded_add, |
837 | info->dram_scrb_add, error_found, handle_error); |
838 | } |
839 | |
840 | static void e752x_get_error_info(struct mem_ctl_info *mci, |
841 | struct e752x_error_info *info) |
842 | { |
843 | struct pci_dev *dev; |
844 | struct e752x_pvt *pvt; |
845 | |
846 | memset(info, 0, sizeof(*info)); |
847 | pvt = (struct e752x_pvt *)mci->pvt_info; |
848 | dev = pvt->dev_d0f1; |
849 | pci_read_config_dword(dev, E752X_FERR_GLOBAL, &info->ferr_global); |
850 | |
851 | if (info->ferr_global) { |
852 | if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) { |
853 | pci_read_config_dword(dev, I3100_NSI_FERR, |
854 | &info->nsi_ferr); |
855 | info->hi_ferr = 0; |
856 | } else { |
857 | pci_read_config_byte(dev, E752X_HI_FERR, |
858 | &info->hi_ferr); |
859 | info->nsi_ferr = 0; |
860 | } |
861 | pci_read_config_word(dev, E752X_SYSBUS_FERR, |
862 | &info->sysbus_ferr); |
863 | pci_read_config_byte(dev, E752X_BUF_FERR, &info->buf_ferr); |
864 | pci_read_config_word(dev, E752X_DRAM_FERR, &info->dram_ferr); |
865 | pci_read_config_dword(dev, E752X_DRAM_SEC1_ADD, |
866 | &info->dram_sec1_add); |
867 | pci_read_config_word(dev, E752X_DRAM_SEC1_SYNDROME, |
868 | &info->dram_sec1_syndrome); |
869 | pci_read_config_dword(dev, E752X_DRAM_DED_ADD, |
870 | &info->dram_ded_add); |
871 | pci_read_config_dword(dev, E752X_DRAM_SCRB_ADD, |
872 | &info->dram_scrb_add); |
873 | pci_read_config_dword(dev, E752X_DRAM_RETR_ADD, |
874 | &info->dram_retr_add); |
875 | |
876 | /* ignore the reserved bits just in case */ |
877 | if (info->hi_ferr & 0x7f) |
878 | pci_write_config_byte(dev, E752X_HI_FERR, |
879 | info->hi_ferr); |
880 | |
881 | if (info->nsi_ferr & NSI_ERR_MASK) |
882 | pci_write_config_dword(dev, I3100_NSI_FERR, |
883 | info->nsi_ferr); |
884 | |
885 | if (info->sysbus_ferr) |
886 | pci_write_config_word(dev, E752X_SYSBUS_FERR, |
887 | info->sysbus_ferr); |
888 | |
889 | if (info->buf_ferr & 0x0f) |
890 | pci_write_config_byte(dev, E752X_BUF_FERR, |
891 | info->buf_ferr); |
892 | |
893 | if (info->dram_ferr) |
894 | pci_write_bits16(pvt->bridge_ck, E752X_DRAM_FERR, |
895 | info->dram_ferr, info->dram_ferr); |
896 | |
897 | pci_write_config_dword(dev, E752X_FERR_GLOBAL, |
898 | info->ferr_global); |
899 | } |
900 | |
901 | pci_read_config_dword(dev, E752X_NERR_GLOBAL, &info->nerr_global); |
902 | |
903 | if (info->nerr_global) { |
904 | if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) { |
905 | pci_read_config_dword(dev, I3100_NSI_NERR, |
906 | &info->nsi_nerr); |
907 | info->hi_nerr = 0; |
908 | } else { |
909 | pci_read_config_byte(dev, E752X_HI_NERR, |
910 | &info->hi_nerr); |
911 | info->nsi_nerr = 0; |
912 | } |
913 | pci_read_config_word(dev, E752X_SYSBUS_NERR, |
914 | &info->sysbus_nerr); |
915 | pci_read_config_byte(dev, E752X_BUF_NERR, &info->buf_nerr); |
916 | pci_read_config_word(dev, E752X_DRAM_NERR, &info->dram_nerr); |
917 | pci_read_config_dword(dev, E752X_DRAM_SEC2_ADD, |
918 | &info->dram_sec2_add); |
919 | pci_read_config_word(dev, E752X_DRAM_SEC2_SYNDROME, |
920 | &info->dram_sec2_syndrome); |
921 | |
922 | if (info->hi_nerr & 0x7f) |
923 | pci_write_config_byte(dev, E752X_HI_NERR, |
924 | info->hi_nerr); |
925 | |
926 | if (info->nsi_nerr & NSI_ERR_MASK) |
927 | pci_write_config_dword(dev, I3100_NSI_NERR, |
928 | info->nsi_nerr); |
929 | |
930 | if (info->sysbus_nerr) |
931 | pci_write_config_word(dev, E752X_SYSBUS_NERR, |
932 | info->sysbus_nerr); |
933 | |
934 | if (info->buf_nerr & 0x0f) |
935 | pci_write_config_byte(dev, E752X_BUF_NERR, |
936 | info->buf_nerr); |
937 | |
938 | if (info->dram_nerr) |
939 | pci_write_bits16(pvt->bridge_ck, E752X_DRAM_NERR, |
940 | info->dram_nerr, info->dram_nerr); |
941 | |
942 | pci_write_config_dword(dev, E752X_NERR_GLOBAL, |
943 | info->nerr_global); |
944 | } |
945 | } |
946 | |
947 | static int e752x_process_error_info(struct mem_ctl_info *mci, |
948 | struct e752x_error_info *info, |
949 | int handle_errors) |
950 | { |
951 | u32 error32, stat32; |
952 | int error_found; |
953 | |
954 | error_found = 0; |
955 | error32 = (info->ferr_global >> 18) & 0x3ff; |
956 | stat32 = (info->ferr_global >> 4) & 0x7ff; |
957 | |
958 | if (error32) |
959 | global_error(1, error32, &error_found, handle_errors); |
960 | |
961 | if (stat32) |
962 | global_error(0, stat32, &error_found, handle_errors); |
963 | |
964 | error32 = (info->nerr_global >> 18) & 0x3ff; |
965 | stat32 = (info->nerr_global >> 4) & 0x7ff; |
966 | |
967 | if (error32) |
968 | global_error(1, error32, &error_found, handle_errors); |
969 | |
970 | if (stat32) |
971 | global_error(0, stat32, &error_found, handle_errors); |
972 | |
973 | e752x_check_hub_interface(info, &error_found, handle_errors); |
974 | e752x_check_ns_interface(info, &error_found, handle_errors); |
975 | e752x_check_sysbus(info, &error_found, handle_errors); |
976 | e752x_check_membuf(info, &error_found, handle_errors); |
977 | e752x_check_dram(mci, info, &error_found, handle_errors); |
978 | return error_found; |
979 | } |
980 | |
981 | static void e752x_check(struct mem_ctl_info *mci) |
982 | { |
983 | struct e752x_error_info info; |
984 | |
985 | edac_dbg(3, "\n"); |
986 | e752x_get_error_info(mci, &info); |
987 | e752x_process_error_info(mci, &info, 1); |
988 | } |
989 | |
990 | /* Program byte/sec bandwidth scrub rate to hardware */ |
991 | static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw) |
992 | { |
993 | const struct scrubrate *scrubrates; |
994 | struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info; |
995 | struct pci_dev *pdev = pvt->dev_d0f0; |
996 | int i; |
997 | |
998 | if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0) |
999 | scrubrates = scrubrates_i3100; |
1000 | else |
1001 | scrubrates = scrubrates_e752x; |
1002 | |
1003 | /* Translate the desired scrub rate to a e752x/3100 register value. |
1004 | * Search for the bandwidth that is equal or greater than the |
1005 | * desired rate and program the cooresponding register value. |
1006 | */ |
1007 | for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++) |
1008 | if (scrubrates[i].bandwidth >= new_bw) |
1009 | break; |
1010 | |
1011 | if (scrubrates[i].bandwidth == SDRATE_EOT) |
1012 | return -1; |
1013 | |
1014 | pci_write_config_word(pdev, E752X_MCHSCRB, scrubrates[i].scrubval); |
1015 | |
1016 | return scrubrates[i].bandwidth; |
1017 | } |
1018 | |
1019 | /* Convert current scrub rate value into byte/sec bandwidth */ |
1020 | static int get_sdram_scrub_rate(struct mem_ctl_info *mci) |
1021 | { |
1022 | const struct scrubrate *scrubrates; |
1023 | struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info; |
1024 | struct pci_dev *pdev = pvt->dev_d0f0; |
1025 | u16 scrubval; |
1026 | int i; |
1027 | |
1028 | if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0) |
1029 | scrubrates = scrubrates_i3100; |
1030 | else |
1031 | scrubrates = scrubrates_e752x; |
1032 | |
1033 | /* Find the bandwidth matching the memory scrubber configuration */ |
1034 | pci_read_config_word(pdev, E752X_MCHSCRB, &scrubval); |
1035 | scrubval = scrubval & 0x0f; |
1036 | |
1037 | for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++) |
1038 | if (scrubrates[i].scrubval == scrubval) |
1039 | break; |
1040 | |
1041 | if (scrubrates[i].bandwidth == SDRATE_EOT) { |
1042 | e752x_printk(KERN_WARNING, |
1043 | "Invalid sdram scrub control value: 0x%x\n", scrubval); |
1044 | return -1; |
1045 | } |
1046 | return scrubrates[i].bandwidth; |
1047 | |
1048 | } |
1049 | |
1050 | /* Return 1 if dual channel mode is active. Else return 0. */ |
1051 | static inline int dual_channel_active(u16 ddrcsr) |
1052 | { |
1053 | return (((ddrcsr >> 12) & 3) == 3); |
1054 | } |
1055 | |
1056 | /* Remap csrow index numbers if map_type is "reverse" |
1057 | */ |
1058 | static inline int remap_csrow_index(struct mem_ctl_info *mci, int index) |
1059 | { |
1060 | struct e752x_pvt *pvt = mci->pvt_info; |
1061 | |
1062 | if (!pvt->map_type) |
1063 | return (7 - index); |
1064 | |
1065 | return (index); |
1066 | } |
1067 | |
1068 | static void e752x_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev, |
1069 | u16 ddrcsr) |
1070 | { |
1071 | struct csrow_info *csrow; |
1072 | enum edac_type edac_mode; |
1073 | unsigned long last_cumul_size; |
1074 | int index, mem_dev, drc_chan; |
1075 | int drc_drbg; /* DRB granularity 0=64mb, 1=128mb */ |
1076 | int drc_ddim; /* DRAM Data Integrity Mode 0=none, 2=edac */ |
1077 | u8 value; |
1078 | u32 dra, drc, cumul_size, i, nr_pages; |
1079 | |
1080 | dra = 0; |
1081 | for (index = 0; index < 4; index++) { |
1082 | u8 dra_reg; |
1083 | pci_read_config_byte(pdev, E752X_DRA + index, &dra_reg); |
1084 | dra |= dra_reg << (index * 8); |
1085 | } |
1086 | pci_read_config_dword(pdev, E752X_DRC, &drc); |
1087 | drc_chan = dual_channel_active(ddrcsr) ? 1 : 0; |
1088 | drc_drbg = drc_chan + 1; /* 128 in dual mode, 64 in single */ |
1089 | drc_ddim = (drc >> 20) & 0x3; |
1090 | |
1091 | /* The dram row boundary (DRB) reg values are boundary address for |
1092 | * each DRAM row with a granularity of 64 or 128MB (single/dual |
1093 | * channel operation). DRB regs are cumulative; therefore DRB7 will |
1094 | * contain the total memory contained in all eight rows. |
1095 | */ |
1096 | for (last_cumul_size = index = 0; index < mci->nr_csrows; index++) { |
1097 | /* mem_dev 0=x8, 1=x4 */ |
1098 | mem_dev = (dra >> (index * 4 + 2)) & 0x3; |
1099 | csrow = mci->csrows[remap_csrow_index(mci, index)]; |
1100 | |
1101 | mem_dev = (mem_dev == 2); |
1102 | pci_read_config_byte(pdev, E752X_DRB + index, &value); |
1103 | /* convert a 128 or 64 MiB DRB to a page size. */ |
1104 | cumul_size = value << (25 + drc_drbg - PAGE_SHIFT); |
1105 | edac_dbg(3, "(%d) cumul_size 0x%x\n", index, cumul_size); |
1106 | if (cumul_size == last_cumul_size) |
1107 | continue; /* not populated */ |
1108 | |
1109 | csrow->first_page = last_cumul_size; |
1110 | csrow->last_page = cumul_size - 1; |
1111 | nr_pages = cumul_size - last_cumul_size; |
1112 | last_cumul_size = cumul_size; |
1113 | |
1114 | /* |
1115 | * if single channel or x8 devices then SECDED |
1116 | * if dual channel and x4 then S4ECD4ED |
1117 | */ |
1118 | if (drc_ddim) { |
1119 | if (drc_chan && mem_dev) { |
1120 | edac_mode = EDAC_S4ECD4ED; |
1121 | mci->edac_cap |= EDAC_FLAG_S4ECD4ED; |
1122 | } else { |
1123 | edac_mode = EDAC_SECDED; |
1124 | mci->edac_cap |= EDAC_FLAG_SECDED; |
1125 | } |
1126 | } else |
1127 | edac_mode = EDAC_NONE; |
1128 | for (i = 0; i < csrow->nr_channels; i++) { |
1129 | struct dimm_info *dimm = csrow->channels[i]->dimm; |
1130 | |
1131 | edac_dbg(3, "Initializing rank at (%i,%i)\n", index, i); |
1132 | dimm->nr_pages = nr_pages / csrow->nr_channels; |
1133 | dimm->grain = 1 << 12; /* 4KiB - resolution of CELOG */ |
1134 | dimm->mtype = MEM_RDDR; /* only one type supported */ |
1135 | dimm->dtype = mem_dev ? DEV_X4 : DEV_X8; |
1136 | dimm->edac_mode = edac_mode; |
1137 | } |
1138 | } |
1139 | } |
1140 | |
1141 | static void e752x_init_mem_map_table(struct pci_dev *pdev, |
1142 | struct e752x_pvt *pvt) |
1143 | { |
1144 | int index; |
1145 | u8 value, last, row; |
1146 | |
1147 | last = 0; |
1148 | row = 0; |
1149 | |
1150 | for (index = 0; index < 8; index += 2) { |
1151 | pci_read_config_byte(pdev, E752X_DRB + index, &value); |
1152 | /* test if there is a dimm in this slot */ |
1153 | if (value == last) { |
1154 | /* no dimm in the slot, so flag it as empty */ |
1155 | pvt->map[index] = 0xff; |
1156 | pvt->map[index + 1] = 0xff; |
1157 | } else { /* there is a dimm in the slot */ |
1158 | pvt->map[index] = row; |
1159 | row++; |
1160 | last = value; |
1161 | /* test the next value to see if the dimm is double |
1162 | * sided |
1163 | */ |
1164 | pci_read_config_byte(pdev, E752X_DRB + index + 1, |
1165 | &value); |
1166 | |
1167 | /* the dimm is single sided, so flag as empty */ |
1168 | /* this is a double sided dimm to save the next row #*/ |
1169 | pvt->map[index + 1] = (value == last) ? 0xff : row; |
1170 | row++; |
1171 | last = value; |
1172 | } |
1173 | } |
1174 | } |
1175 | |
1176 | /* Return 0 on success or 1 on failure. */ |
1177 | static int e752x_get_devs(struct pci_dev *pdev, int dev_idx, |
1178 | struct e752x_pvt *pvt) |
1179 | { |
1180 | struct pci_dev *dev; |
1181 | |
1182 | pvt->bridge_ck = pci_get_device(PCI_VENDOR_ID_INTEL, |
1183 | pvt->dev_info->err_dev, pvt->bridge_ck); |
1184 | |
1185 | if (pvt->bridge_ck == NULL) |
1186 | pvt->bridge_ck = pci_scan_single_device(pdev->bus, |
1187 | PCI_DEVFN(0, 1)); |
1188 | |
1189 | if (pvt->bridge_ck == NULL) { |
1190 | e752x_printk(KERN_ERR, "error reporting device not found:" |
1191 | "vendor %x device 0x%x (broken BIOS?)\n", |
1192 | PCI_VENDOR_ID_INTEL, e752x_devs[dev_idx].err_dev); |
1193 | return 1; |
1194 | } |
1195 | |
1196 | dev = pci_get_device(PCI_VENDOR_ID_INTEL, |
1197 | e752x_devs[dev_idx].ctl_dev, |
1198 | NULL); |
1199 | |
1200 | if (dev == NULL) |
1201 | goto fail; |
1202 | |
1203 | pvt->dev_d0f0 = dev; |
1204 | pvt->dev_d0f1 = pci_dev_get(pvt->bridge_ck); |
1205 | |
1206 | return 0; |
1207 | |
1208 | fail: |
1209 | pci_dev_put(pvt->bridge_ck); |
1210 | return 1; |
1211 | } |
1212 | |
1213 | /* Setup system bus parity mask register. |
1214 | * Sysbus parity supported on: |
1215 | * e7320/e7520/e7525 + Xeon |
1216 | */ |
1217 | static void e752x_init_sysbus_parity_mask(struct e752x_pvt *pvt) |
1218 | { |
1219 | char *cpu_id = cpu_data(0).x86_model_id; |
1220 | struct pci_dev *dev = pvt->dev_d0f1; |
1221 | int enable = 1; |
1222 | |
1223 | /* Allow module parameter override, else see if CPU supports parity */ |
1224 | if (sysbus_parity != -1) { |
1225 | enable = sysbus_parity; |
1226 | } else if (cpu_id[0] && !strstr(cpu_id, "Xeon")) { |
1227 | e752x_printk(KERN_INFO, "System Bus Parity not " |
1228 | "supported by CPU, disabling\n"); |
1229 | enable = 0; |
1230 | } |
1231 | |
1232 | if (enable) |
1233 | pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0000); |
1234 | else |
1235 | pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0309); |
1236 | } |
1237 | |
1238 | static void e752x_init_error_reporting_regs(struct e752x_pvt *pvt) |
1239 | { |
1240 | struct pci_dev *dev; |
1241 | |
1242 | dev = pvt->dev_d0f1; |
1243 | /* Turn off error disable & SMI in case the BIOS turned it on */ |
1244 | if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) { |
1245 | pci_write_config_dword(dev, I3100_NSI_EMASK, 0); |
1246 | pci_write_config_dword(dev, I3100_NSI_SMICMD, 0); |
1247 | } else { |
1248 | pci_write_config_byte(dev, E752X_HI_ERRMASK, 0x00); |
1249 | pci_write_config_byte(dev, E752X_HI_SMICMD, 0x00); |
1250 | } |
1251 | |
1252 | e752x_init_sysbus_parity_mask(pvt); |
1253 | |
1254 | pci_write_config_word(dev, E752X_SYSBUS_SMICMD, 0x00); |
1255 | pci_write_config_byte(dev, E752X_BUF_ERRMASK, 0x00); |
1256 | pci_write_config_byte(dev, E752X_BUF_SMICMD, 0x00); |
1257 | pci_write_config_byte(dev, E752X_DRAM_ERRMASK, 0x00); |
1258 | pci_write_config_byte(dev, E752X_DRAM_SMICMD, 0x00); |
1259 | } |
1260 | |
1261 | static int e752x_probe1(struct pci_dev *pdev, int dev_idx) |
1262 | { |
1263 | u16 pci_data; |
1264 | u8 stat8; |
1265 | struct mem_ctl_info *mci; |
1266 | struct edac_mc_layer layers[2]; |
1267 | struct e752x_pvt *pvt; |
1268 | u16 ddrcsr; |
1269 | int drc_chan; /* Number of channels 0=1chan,1=2chan */ |
1270 | struct e752x_error_info discard; |
1271 | |
1272 | edac_dbg(0, "mci\n"); |
1273 | edac_dbg(0, "Starting Probe1\n"); |
1274 | |
1275 | /* check to see if device 0 function 1 is enabled; if it isn't, we |
1276 | * assume the BIOS has reserved it for a reason and is expecting |
1277 | * exclusive access, we take care not to violate that assumption and |
1278 | * fail the probe. */ |
1279 | pci_read_config_byte(pdev, E752X_DEVPRES1, &stat8); |
1280 | if (!force_function_unhide && !(stat8 & (1 << 5))) { |
1281 | printk(KERN_INFO "Contact your BIOS vendor to see if the " |
1282 | "E752x error registers can be safely un-hidden\n"); |
1283 | return -ENODEV; |
1284 | } |
1285 | stat8 |= (1 << 5); |
1286 | pci_write_config_byte(pdev, E752X_DEVPRES1, stat8); |
1287 | |
1288 | pci_read_config_word(pdev, E752X_DDRCSR, &ddrcsr); |
1289 | /* FIXME: should check >>12 or 0xf, true for all? */ |
1290 | /* Dual channel = 1, Single channel = 0 */ |
1291 | drc_chan = dual_channel_active(ddrcsr); |
1292 | |
1293 | layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; |
1294 | layers[0].size = E752X_NR_CSROWS; |
1295 | layers[0].is_virt_csrow = true; |
1296 | layers[1].type = EDAC_MC_LAYER_CHANNEL; |
1297 | layers[1].size = drc_chan + 1; |
1298 | layers[1].is_virt_csrow = false; |
1299 | mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt)); |
1300 | if (mci == NULL) |
1301 | return -ENOMEM; |
1302 | |
1303 | edac_dbg(3, "init mci\n"); |
1304 | mci->mtype_cap = MEM_FLAG_RDDR; |
1305 | /* 3100 IMCH supports SECDEC only */ |
1306 | mci->edac_ctl_cap = (dev_idx == I3100) ? EDAC_FLAG_SECDED : |
1307 | (EDAC_FLAG_NONE | EDAC_FLAG_SECDED | EDAC_FLAG_S4ECD4ED); |
1308 | /* FIXME - what if different memory types are in different csrows? */ |
1309 | mci->mod_name = EDAC_MOD_STR; |
1310 | mci->mod_ver = E752X_REVISION; |
1311 | mci->pdev = &pdev->dev; |
1312 | |
1313 | edac_dbg(3, "init pvt\n"); |
1314 | pvt = (struct e752x_pvt *)mci->pvt_info; |
1315 | pvt->dev_info = &e752x_devs[dev_idx]; |
1316 | pvt->mc_symmetric = ((ddrcsr & 0x10) != 0); |
1317 | |
1318 | if (e752x_get_devs(pdev, dev_idx, pvt)) { |
1319 | edac_mc_free(mci); |
1320 | return -ENODEV; |
1321 | } |
1322 | |
1323 | edac_dbg(3, "more mci init\n"); |
1324 | mci->ctl_name = pvt->dev_info->ctl_name; |
1325 | mci->dev_name = pci_name(pdev); |
1326 | mci->edac_check = e752x_check; |
1327 | mci->ctl_page_to_phys = ctl_page_to_phys; |
1328 | mci->set_sdram_scrub_rate = set_sdram_scrub_rate; |
1329 | mci->get_sdram_scrub_rate = get_sdram_scrub_rate; |
1330 | |
1331 | /* set the map type. 1 = normal, 0 = reversed |
1332 | * Must be set before e752x_init_csrows in case csrow mapping |
1333 | * is reversed. |
1334 | */ |
1335 | pci_read_config_byte(pdev, E752X_DRM, &stat8); |
1336 | pvt->map_type = ((stat8 & 0x0f) > ((stat8 >> 4) & 0x0f)); |
1337 | |
1338 | e752x_init_csrows(mci, pdev, ddrcsr); |
1339 | e752x_init_mem_map_table(pdev, pvt); |
1340 | |
1341 | if (dev_idx == I3100) |
1342 | mci->edac_cap = EDAC_FLAG_SECDED; /* the only mode supported */ |
1343 | else |
1344 | mci->edac_cap |= EDAC_FLAG_NONE; |
1345 | edac_dbg(3, "tolm, remapbase, remaplimit\n"); |
1346 | |
1347 | /* load the top of low memory, remap base, and remap limit vars */ |
1348 | pci_read_config_word(pdev, E752X_TOLM, &pci_data); |
1349 | pvt->tolm = ((u32) pci_data) << 4; |
1350 | pci_read_config_word(pdev, E752X_REMAPBASE, &pci_data); |
1351 | pvt->remapbase = ((u32) pci_data) << 14; |
1352 | pci_read_config_word(pdev, E752X_REMAPLIMIT, &pci_data); |
1353 | pvt->remaplimit = ((u32) pci_data) << 14; |
1354 | e752x_printk(KERN_INFO, |
1355 | "tolm = %x, remapbase = %x, remaplimit = %x\n", |
1356 | pvt->tolm, pvt->remapbase, pvt->remaplimit); |
1357 | |
1358 | /* Here we assume that we will never see multiple instances of this |
1359 | * type of memory controller. The ID is therefore hardcoded to 0. |
1360 | */ |
1361 | if (edac_mc_add_mc(mci)) { |
1362 | edac_dbg(3, "failed edac_mc_add_mc()\n"); |
1363 | goto fail; |
1364 | } |
1365 | |
1366 | e752x_init_error_reporting_regs(pvt); |
1367 | e752x_get_error_info(mci, &discard); /* clear other MCH errors */ |
1368 | |
1369 | /* allocating generic PCI control info */ |
1370 | e752x_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR); |
1371 | if (!e752x_pci) { |
1372 | printk(KERN_WARNING |
1373 | "%s(): Unable to create PCI control\n", __func__); |
1374 | printk(KERN_WARNING |
1375 | "%s(): PCI error report via EDAC not setup\n", |
1376 | __func__); |
1377 | } |
1378 | |
1379 | /* get this far and it's successful */ |
1380 | edac_dbg(3, "success\n"); |
1381 | return 0; |
1382 | |
1383 | fail: |
1384 | pci_dev_put(pvt->dev_d0f0); |
1385 | pci_dev_put(pvt->dev_d0f1); |
1386 | pci_dev_put(pvt->bridge_ck); |
1387 | edac_mc_free(mci); |
1388 | |
1389 | return -ENODEV; |
1390 | } |
1391 | |
1392 | /* returns count (>= 0), or negative on error */ |
1393 | static int __devinit e752x_init_one(struct pci_dev *pdev, |
1394 | const struct pci_device_id *ent) |
1395 | { |
1396 | edac_dbg(0, "\n"); |
1397 | |
1398 | /* wake up and enable device */ |
1399 | if (pci_enable_device(pdev) < 0) |
1400 | return -EIO; |
1401 | |
1402 | return e752x_probe1(pdev, ent->driver_data); |
1403 | } |
1404 | |
1405 | static void __devexit e752x_remove_one(struct pci_dev *pdev) |
1406 | { |
1407 | struct mem_ctl_info *mci; |
1408 | struct e752x_pvt *pvt; |
1409 | |
1410 | edac_dbg(0, "\n"); |
1411 | |
1412 | if (e752x_pci) |
1413 | edac_pci_release_generic_ctl(e752x_pci); |
1414 | |
1415 | if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL) |
1416 | return; |
1417 | |
1418 | pvt = (struct e752x_pvt *)mci->pvt_info; |
1419 | pci_dev_put(pvt->dev_d0f0); |
1420 | pci_dev_put(pvt->dev_d0f1); |
1421 | pci_dev_put(pvt->bridge_ck); |
1422 | edac_mc_free(mci); |
1423 | } |
1424 | |
1425 | static DEFINE_PCI_DEVICE_TABLE(e752x_pci_tbl) = { |
1426 | { |
1427 | PCI_VEND_DEV(INTEL, 7520_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, |
1428 | E7520}, |
1429 | { |
1430 | PCI_VEND_DEV(INTEL, 7525_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, |
1431 | E7525}, |
1432 | { |
1433 | PCI_VEND_DEV(INTEL, 7320_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, |
1434 | E7320}, |
1435 | { |
1436 | PCI_VEND_DEV(INTEL, 3100_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, |
1437 | I3100}, |
1438 | { |
1439 | 0, |
1440 | } /* 0 terminated list. */ |
1441 | }; |
1442 | |
1443 | MODULE_DEVICE_TABLE(pci, e752x_pci_tbl); |
1444 | |
1445 | static struct pci_driver e752x_driver = { |
1446 | .name = EDAC_MOD_STR, |
1447 | .probe = e752x_init_one, |
1448 | .remove = __devexit_p(e752x_remove_one), |
1449 | .id_table = e752x_pci_tbl, |
1450 | }; |
1451 | |
1452 | static int __init e752x_init(void) |
1453 | { |
1454 | int pci_rc; |
1455 | |
1456 | edac_dbg(3, "\n"); |
1457 | |
1458 | /* Ensure that the OPSTATE is set correctly for POLL or NMI */ |
1459 | opstate_init(); |
1460 | |
1461 | pci_rc = pci_register_driver(&e752x_driver); |
1462 | return (pci_rc < 0) ? pci_rc : 0; |
1463 | } |
1464 | |
1465 | static void __exit e752x_exit(void) |
1466 | { |
1467 | edac_dbg(3, "\n"); |
1468 | pci_unregister_driver(&e752x_driver); |
1469 | } |
1470 | |
1471 | module_init(e752x_init); |
1472 | module_exit(e752x_exit); |
1473 | |
1474 | MODULE_LICENSE("GPL"); |
1475 | MODULE_AUTHOR("Linux Networx (http://lnxi.com) Tom Zimmerman\n"); |
1476 | MODULE_DESCRIPTION("MC support for Intel e752x/3100 memory controllers"); |
1477 | |
1478 | module_param(force_function_unhide, int, 0444); |
1479 | MODULE_PARM_DESC(force_function_unhide, "if BIOS sets Dev0:Fun1 up as hidden:" |
1480 | " 1=force unhide and hope BIOS doesn't fight driver for " |
1481 | "Dev0:Fun1 access"); |
1482 | |
1483 | module_param(edac_op_state, int, 0444); |
1484 | MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI"); |
1485 | |
1486 | module_param(sysbus_parity, int, 0444); |
1487 | MODULE_PARM_DESC(sysbus_parity, "0=disable system bus parity checking," |
1488 | " 1=enable system bus parity checking, default=auto-detect"); |
1489 | module_param(report_non_memory_errors, int, 0644); |
1490 | MODULE_PARM_DESC(report_non_memory_errors, "0=disable non-memory error " |
1491 | "reporting, 1=enable non-memory error reporting"); |
1492 |
Branches:
ben-wpan
ben-wpan-stefan
javiroman/ks7010
jz-2.6.34
jz-2.6.34-rc5
jz-2.6.34-rc6
jz-2.6.34-rc7
jz-2.6.35
jz-2.6.36
jz-2.6.37
jz-2.6.38
jz-2.6.39
jz-3.0
jz-3.1
jz-3.11
jz-3.12
jz-3.13
jz-3.15
jz-3.16
jz-3.18-dt
jz-3.2
jz-3.3
jz-3.4
jz-3.5
jz-3.6
jz-3.6-rc2-pwm
jz-3.9
jz-3.9-clk
jz-3.9-rc8
jz47xx
jz47xx-2.6.38
master
Tags:
od-2011-09-04
od-2011-09-18
v2.6.34-rc5
v2.6.34-rc6
v2.6.34-rc7
v3.9