]> git.proxmox.com Git - mirror_ubuntu-eoan-kernel.git/blame - drivers/edac/amd64_edac.c
UBUNTU: Ubuntu-5.3.0-29.31
[mirror_ubuntu-eoan-kernel.git] / drivers / edac / amd64_edac.c
CommitLineData
09c434b8 1// SPDX-License-Identifier: GPL-2.0-only
2bc65418 2#include "amd64_edac.h"
23ac4ae8 3#include <asm/amd_nb.h>
2bc65418 4
d1ea71cd 5static struct edac_pci_ctl_info *pci_ctl;
2bc65418
DT
6
7static int report_gart_errors;
8module_param(report_gart_errors, int, 0644);
9
10/*
11 * Set by command line parameter. If BIOS has enabled the ECC, this override is
12 * cleared to prevent re-enabling the hardware by this driver.
13 */
14static int ecc_enable_override;
15module_param(ecc_enable_override, int, 0644);
16
a29d8b8e 17static struct msr __percpu *msrs;
50542251 18
2ec591ac 19/* Per-node stuff */
ae7bb7c6 20static struct ecc_settings **ecc_stngs;
2bc65418 21
bdcee774
YG
22/* Number of Unified Memory Controllers */
23static u8 num_umcs;
24
b70ef010
BP
25/*
26 * Valid scrub rates for the K8 hardware memory scrubber. We map the scrubbing
27 * bandwidth to a valid bit pattern. The 'set' operation finds the 'matching-
28 * or higher value'.
29 *
30 *FIXME: Produce a better mapping/linearisation.
31 */
c7e5301a 32static const struct scrubrate {
39094443
BP
33 u32 scrubval; /* bit pattern for scrub rate */
34 u32 bandwidth; /* bandwidth consumed (bytes/sec) */
35} scrubrates[] = {
b70ef010
BP
36 { 0x01, 1600000000UL},
37 { 0x02, 800000000UL},
38 { 0x03, 400000000UL},
39 { 0x04, 200000000UL},
40 { 0x05, 100000000UL},
41 { 0x06, 50000000UL},
42 { 0x07, 25000000UL},
43 { 0x08, 12284069UL},
44 { 0x09, 6274509UL},
45 { 0x0A, 3121951UL},
46 { 0x0B, 1560975UL},
47 { 0x0C, 781440UL},
48 { 0x0D, 390720UL},
49 { 0x0E, 195300UL},
50 { 0x0F, 97650UL},
51 { 0x10, 48854UL},
52 { 0x11, 24427UL},
53 { 0x12, 12213UL},
54 { 0x13, 6101UL},
55 { 0x14, 3051UL},
56 { 0x15, 1523UL},
57 { 0x16, 761UL},
58 { 0x00, 0UL}, /* scrubbing off */
59};
60
66fed2d4
BP
61int __amd64_read_pci_cfg_dword(struct pci_dev *pdev, int offset,
62 u32 *val, const char *func)
b2b0c605
BP
63{
64 int err = 0;
65
66 err = pci_read_config_dword(pdev, offset, val);
67 if (err)
68 amd64_warn("%s: error reading F%dx%03x.\n",
69 func, PCI_FUNC(pdev->devfn), offset);
70
71 return err;
72}
73
74int __amd64_write_pci_cfg_dword(struct pci_dev *pdev, int offset,
75 u32 val, const char *func)
76{
77 int err = 0;
78
79 err = pci_write_config_dword(pdev, offset, val);
80 if (err)
81 amd64_warn("%s: error writing to F%dx%03x.\n",
82 func, PCI_FUNC(pdev->devfn), offset);
83
84 return err;
85}
86
7981a28f
AG
87/*
88 * Select DCT to which PCI cfg accesses are routed
89 */
90static void f15h_select_dct(struct amd64_pvt *pvt, u8 dct)
91{
92 u32 reg = 0;
93
94 amd64_read_pci_cfg(pvt->F1, DCT_CFG_SEL, &reg);
95 reg &= (pvt->model == 0x30) ? ~3 : ~1;
96 reg |= dct;
97 amd64_write_pci_cfg(pvt->F1, DCT_CFG_SEL, reg);
98}
99
b2b0c605
BP
100/*
101 *
102 * Depending on the family, F2 DCT reads need special handling:
103 *
7981a28f 104 * K8: has a single DCT only and no address offsets >= 0x100
b2b0c605
BP
105 *
106 * F10h: each DCT has its own set of regs
107 * DCT0 -> F2x040..
108 * DCT1 -> F2x140..
109 *
94c1acf2 110 * F16h: has only 1 DCT
7981a28f
AG
111 *
112 * F15h: we select which DCT we access using F1x10C[DctCfgSel]
b2b0c605 113 */
7981a28f
AG
114static inline int amd64_read_dct_pci_cfg(struct amd64_pvt *pvt, u8 dct,
115 int offset, u32 *val)
b2b0c605 116{
7981a28f
AG
117 switch (pvt->fam) {
118 case 0xf:
119 if (dct || offset >= 0x100)
120 return -EINVAL;
121 break;
b2b0c605 122
7981a28f
AG
123 case 0x10:
124 if (dct) {
125 /*
126 * Note: If ganging is enabled, barring the regs
127 * F2x[1,0]98 and F2x[1,0]9C; reads reads to F2x1xx
128 * return 0. (cf. Section 2.8.1 F10h BKDG)
129 */
130 if (dct_ganging_enabled(pvt))
131 return 0;
b2b0c605 132
7981a28f
AG
133 offset += 0x100;
134 }
135 break;
73ba8593 136
7981a28f
AG
137 case 0x15:
138 /*
139 * F15h: F2x1xx addresses do not map explicitly to DCT1.
140 * We should select which DCT we access using F1x10C[DctCfgSel]
141 */
142 dct = (dct && pvt->model == 0x30) ? 3 : dct;
143 f15h_select_dct(pvt, dct);
144 break;
73ba8593 145
7981a28f
AG
146 case 0x16:
147 if (dct)
148 return -EINVAL;
149 break;
b2b0c605 150
7981a28f
AG
151 default:
152 break;
b2b0c605 153 }
7981a28f 154 return amd64_read_pci_cfg(pvt->F2, offset, val);
b2b0c605
BP
155}
156
2bc65418
DT
157/*
158 * Memory scrubber control interface. For K8, memory scrubbing is handled by
159 * hardware and can involve L2 cache, dcache as well as the main memory. With
160 * F10, this is extended to L3 cache scrubbing on CPU models sporting that
161 * functionality.
162 *
163 * This causes the "units" for the scrubbing speed to vary from 64 byte blocks
164 * (dram) over to cache lines. This is nasty, so we will use bandwidth in
165 * bytes/sec for the setting.
166 *
167 * Currently, we only do dram scrubbing. If the scrubbing is done in software on
168 * other archs, we might not have access to the caches directly.
169 */
170
8051c0af
YG
171static inline void __f17h_set_scrubval(struct amd64_pvt *pvt, u32 scrubval)
172{
173 /*
174 * Fam17h supports scrub values between 0x5 and 0x14. Also, the values
175 * are shifted down by 0x5, so scrubval 0x5 is written to the register
176 * as 0x0, scrubval 0x6 as 0x1, etc.
177 */
178 if (scrubval >= 0x5 && scrubval <= 0x14) {
179 scrubval -= 0x5;
180 pci_write_bits32(pvt->F6, F17H_SCR_LIMIT_ADDR, scrubval, 0xF);
181 pci_write_bits32(pvt->F6, F17H_SCR_BASE_ADDR, 1, 0x1);
182 } else {
183 pci_write_bits32(pvt->F6, F17H_SCR_BASE_ADDR, 0, 0x1);
184 }
185}
2bc65418 186/*
8051c0af 187 * Scan the scrub rate mapping table for a close or matching bandwidth value to
2bc65418
DT
188 * issue. If requested is too big, then use last maximum value found.
189 */
da92110d 190static int __set_scrub_rate(struct amd64_pvt *pvt, u32 new_bw, u32 min_rate)
2bc65418
DT
191{
192 u32 scrubval;
193 int i;
194
195 /*
196 * map the configured rate (new_bw) to a value specific to the AMD64
197 * memory controller and apply to register. Search for the first
198 * bandwidth entry that is greater or equal than the setting requested
199 * and program that. If at last entry, turn off DRAM scrubbing.
168bfeef
AM
200 *
201 * If no suitable bandwidth is found, turn off DRAM scrubbing entirely
202 * by falling back to the last element in scrubrates[].
2bc65418 203 */
168bfeef 204 for (i = 0; i < ARRAY_SIZE(scrubrates) - 1; i++) {
2bc65418
DT
205 /*
206 * skip scrub rates which aren't recommended
207 * (see F10 BKDG, F3x58)
208 */
395ae783 209 if (scrubrates[i].scrubval < min_rate)
2bc65418
DT
210 continue;
211
212 if (scrubrates[i].bandwidth <= new_bw)
213 break;
2bc65418
DT
214 }
215
216 scrubval = scrubrates[i].scrubval;
2bc65418 217
c4a3e946 218 if (pvt->fam == 0x17 || pvt->fam == 0x18) {
8051c0af
YG
219 __f17h_set_scrubval(pvt, scrubval);
220 } else if (pvt->fam == 0x15 && pvt->model == 0x60) {
da92110d
AG
221 f15h_select_dct(pvt, 0);
222 pci_write_bits32(pvt->F2, F15H_M60H_SCRCTRL, scrubval, 0x001F);
223 f15h_select_dct(pvt, 1);
224 pci_write_bits32(pvt->F2, F15H_M60H_SCRCTRL, scrubval, 0x001F);
225 } else {
226 pci_write_bits32(pvt->F3, SCRCTRL, scrubval, 0x001F);
227 }
2bc65418 228
39094443
BP
229 if (scrubval)
230 return scrubrates[i].bandwidth;
231
2bc65418
DT
232 return 0;
233}
234
d1ea71cd 235static int set_scrub_rate(struct mem_ctl_info *mci, u32 bw)
2bc65418
DT
236{
237 struct amd64_pvt *pvt = mci->pvt_info;
87b3e0e6 238 u32 min_scrubrate = 0x5;
2bc65418 239
a4b4bedc 240 if (pvt->fam == 0xf)
87b3e0e6
BP
241 min_scrubrate = 0x0;
242
da92110d
AG
243 if (pvt->fam == 0x15) {
244 /* Erratum #505 */
245 if (pvt->model < 0x10)
246 f15h_select_dct(pvt, 0);
73ba8593 247
da92110d
AG
248 if (pvt->model == 0x60)
249 min_scrubrate = 0x6;
250 }
251 return __set_scrub_rate(pvt, bw, min_scrubrate);
2bc65418
DT
252}
253
d1ea71cd 254static int get_scrub_rate(struct mem_ctl_info *mci)
2bc65418
DT
255{
256 struct amd64_pvt *pvt = mci->pvt_info;
39094443 257 int i, retval = -EINVAL;
8051c0af 258 u32 scrubval = 0;
2bc65418 259
8051c0af
YG
260 switch (pvt->fam) {
261 case 0x15:
da92110d
AG
262 /* Erratum #505 */
263 if (pvt->model < 0x10)
264 f15h_select_dct(pvt, 0);
73ba8593 265
da92110d
AG
266 if (pvt->model == 0x60)
267 amd64_read_pci_cfg(pvt->F2, F15H_M60H_SCRCTRL, &scrubval);
8051c0af
YG
268 break;
269
270 case 0x17:
c4a3e946 271 case 0x18:
8051c0af
YG
272 amd64_read_pci_cfg(pvt->F6, F17H_SCR_BASE_ADDR, &scrubval);
273 if (scrubval & BIT(0)) {
274 amd64_read_pci_cfg(pvt->F6, F17H_SCR_LIMIT_ADDR, &scrubval);
275 scrubval &= 0xF;
276 scrubval += 0x5;
277 } else {
278 scrubval = 0;
279 }
280 break;
281
282 default:
da92110d 283 amd64_read_pci_cfg(pvt->F3, SCRCTRL, &scrubval);
8051c0af
YG
284 break;
285 }
2bc65418
DT
286
287 scrubval = scrubval & 0x001F;
288
926311fd 289 for (i = 0; i < ARRAY_SIZE(scrubrates); i++) {
2bc65418 290 if (scrubrates[i].scrubval == scrubval) {
39094443 291 retval = scrubrates[i].bandwidth;
2bc65418
DT
292 break;
293 }
294 }
39094443 295 return retval;
2bc65418
DT
296}
297
6775763a 298/*
7f19bf75
BP
299 * returns true if the SysAddr given by sys_addr matches the
300 * DRAM base/limit associated with node_id
6775763a 301 */
d1ea71cd 302static bool base_limit_match(struct amd64_pvt *pvt, u64 sys_addr, u8 nid)
6775763a 303{
7f19bf75 304 u64 addr;
6775763a
DT
305
306 /* The K8 treats this as a 40-bit value. However, bits 63-40 will be
307 * all ones if the most significant implemented address bit is 1.
308 * Here we discard bits 63-40. See section 3.4.2 of AMD publication
309 * 24592: AMD x86-64 Architecture Programmer's Manual Volume 1
310 * Application Programming.
311 */
312 addr = sys_addr & 0x000000ffffffffffull;
313
7f19bf75
BP
314 return ((addr >= get_dram_base(pvt, nid)) &&
315 (addr <= get_dram_limit(pvt, nid)));
6775763a
DT
316}
317
318/*
319 * Attempt to map a SysAddr to a node. On success, return a pointer to the
320 * mem_ctl_info structure for the node that the SysAddr maps to.
321 *
322 * On failure, return NULL.
323 */
324static struct mem_ctl_info *find_mc_by_sys_addr(struct mem_ctl_info *mci,
325 u64 sys_addr)
326{
327 struct amd64_pvt *pvt;
c7e5301a 328 u8 node_id;
6775763a
DT
329 u32 intlv_en, bits;
330
331 /*
332 * Here we use the DRAM Base (section 3.4.4.1) and DRAM Limit (section
333 * 3.4.4.2) registers to map the SysAddr to a node ID.
334 */
335 pvt = mci->pvt_info;
336
337 /*
338 * The value of this field should be the same for all DRAM Base
339 * registers. Therefore we arbitrarily choose to read it from the
340 * register for node 0.
341 */
7f19bf75 342 intlv_en = dram_intlv_en(pvt, 0);
6775763a
DT
343
344 if (intlv_en == 0) {
7f19bf75 345 for (node_id = 0; node_id < DRAM_RANGES; node_id++) {
d1ea71cd 346 if (base_limit_match(pvt, sys_addr, node_id))
8edc5445 347 goto found;
6775763a 348 }
8edc5445 349 goto err_no_match;
6775763a
DT
350 }
351
72f158fe
BP
352 if (unlikely((intlv_en != 0x01) &&
353 (intlv_en != 0x03) &&
354 (intlv_en != 0x07))) {
24f9a7fe 355 amd64_warn("DRAM Base[IntlvEn] junk value: 0x%x, BIOS bug?\n", intlv_en);
6775763a
DT
356 return NULL;
357 }
358
359 bits = (((u32) sys_addr) >> 12) & intlv_en;
360
361 for (node_id = 0; ; ) {
7f19bf75 362 if ((dram_intlv_sel(pvt, node_id) & intlv_en) == bits)
6775763a
DT
363 break; /* intlv_sel field matches */
364
7f19bf75 365 if (++node_id >= DRAM_RANGES)
6775763a
DT
366 goto err_no_match;
367 }
368
369 /* sanity test for sys_addr */
d1ea71cd 370 if (unlikely(!base_limit_match(pvt, sys_addr, node_id))) {
24f9a7fe
BP
371 amd64_warn("%s: sys_addr 0x%llx falls outside base/limit address"
372 "range for node %d with node interleaving enabled.\n",
373 __func__, sys_addr, node_id);
6775763a
DT
374 return NULL;
375 }
376
377found:
b487c33e 378 return edac_mc_find((int)node_id);
6775763a
DT
379
380err_no_match:
956b9ba1
JP
381 edac_dbg(2, "sys_addr 0x%lx doesn't match any node\n",
382 (unsigned long)sys_addr);
6775763a
DT
383
384 return NULL;
385}
e2ce7255
DT
386
387/*
11c75ead
BP
388 * compute the CS base address of the @csrow on the DRAM controller @dct.
389 * For details see F2x[5C:40] in the processor's BKDG
e2ce7255 390 */
11c75ead
BP
391static void get_cs_base_and_mask(struct amd64_pvt *pvt, int csrow, u8 dct,
392 u64 *base, u64 *mask)
e2ce7255 393{
11c75ead
BP
394 u64 csbase, csmask, base_bits, mask_bits;
395 u8 addr_shift;
e2ce7255 396
18b94f66 397 if (pvt->fam == 0xf && pvt->ext_model < K8_REV_F) {
11c75ead
BP
398 csbase = pvt->csels[dct].csbases[csrow];
399 csmask = pvt->csels[dct].csmasks[csrow];
10ef6b0d
CG
400 base_bits = GENMASK_ULL(31, 21) | GENMASK_ULL(15, 9);
401 mask_bits = GENMASK_ULL(29, 21) | GENMASK_ULL(15, 9);
11c75ead 402 addr_shift = 4;
94c1acf2
AG
403
404 /*
18b94f66
AG
405 * F16h and F15h, models 30h and later need two addr_shift values:
406 * 8 for high and 6 for low (cf. F16h BKDG).
407 */
408 } else if (pvt->fam == 0x16 ||
409 (pvt->fam == 0x15 && pvt->model >= 0x30)) {
94c1acf2
AG
410 csbase = pvt->csels[dct].csbases[csrow];
411 csmask = pvt->csels[dct].csmasks[csrow >> 1];
412
10ef6b0d
CG
413 *base = (csbase & GENMASK_ULL(15, 5)) << 6;
414 *base |= (csbase & GENMASK_ULL(30, 19)) << 8;
94c1acf2
AG
415
416 *mask = ~0ULL;
417 /* poke holes for the csmask */
10ef6b0d
CG
418 *mask &= ~((GENMASK_ULL(15, 5) << 6) |
419 (GENMASK_ULL(30, 19) << 8));
94c1acf2 420
10ef6b0d
CG
421 *mask |= (csmask & GENMASK_ULL(15, 5)) << 6;
422 *mask |= (csmask & GENMASK_ULL(30, 19)) << 8;
94c1acf2
AG
423
424 return;
11c75ead
BP
425 } else {
426 csbase = pvt->csels[dct].csbases[csrow];
427 csmask = pvt->csels[dct].csmasks[csrow >> 1];
428 addr_shift = 8;
e2ce7255 429
a4b4bedc 430 if (pvt->fam == 0x15)
10ef6b0d
CG
431 base_bits = mask_bits =
432 GENMASK_ULL(30,19) | GENMASK_ULL(13,5);
11c75ead 433 else
10ef6b0d
CG
434 base_bits = mask_bits =
435 GENMASK_ULL(28,19) | GENMASK_ULL(13,5);
11c75ead 436 }
e2ce7255 437
11c75ead 438 *base = (csbase & base_bits) << addr_shift;
e2ce7255 439
11c75ead
BP
440 *mask = ~0ULL;
441 /* poke holes for the csmask */
442 *mask &= ~(mask_bits << addr_shift);
443 /* OR them in */
444 *mask |= (csmask & mask_bits) << addr_shift;
e2ce7255
DT
445}
446
11c75ead
BP
447#define for_each_chip_select(i, dct, pvt) \
448 for (i = 0; i < pvt->csels[dct].b_cnt; i++)
449
614ec9d8
BP
450#define chip_select_base(i, dct, pvt) \
451 pvt->csels[dct].csbases[i]
452
11c75ead
BP
453#define for_each_chip_select_mask(i, dct, pvt) \
454 for (i = 0; i < pvt->csels[dct].m_cnt; i++)
455
4d30d2bc 456#define for_each_umc(i) \
bdcee774 457 for (i = 0; i < num_umcs; i++)
4d30d2bc 458
e2ce7255
DT
459/*
460 * @input_addr is an InputAddr associated with the node given by mci. Return the
461 * csrow that input_addr maps to, or -1 on failure (no csrow claims input_addr).
462 */
463static int input_addr_to_csrow(struct mem_ctl_info *mci, u64 input_addr)
464{
465 struct amd64_pvt *pvt;
466 int csrow;
467 u64 base, mask;
468
469 pvt = mci->pvt_info;
470
11c75ead
BP
471 for_each_chip_select(csrow, 0, pvt) {
472 if (!csrow_enabled(csrow, 0, pvt))
e2ce7255
DT
473 continue;
474
11c75ead
BP
475 get_cs_base_and_mask(pvt, csrow, 0, &base, &mask);
476
477 mask = ~mask;
e2ce7255
DT
478
479 if ((input_addr & mask) == (base & mask)) {
956b9ba1
JP
480 edac_dbg(2, "InputAddr 0x%lx matches csrow %d (node %d)\n",
481 (unsigned long)input_addr, csrow,
482 pvt->mc_node_id);
e2ce7255
DT
483
484 return csrow;
485 }
486 }
956b9ba1
JP
487 edac_dbg(2, "no matching csrow for InputAddr 0x%lx (MC node %d)\n",
488 (unsigned long)input_addr, pvt->mc_node_id);
e2ce7255
DT
489
490 return -1;
491}
492
e2ce7255
DT
493/*
494 * Obtain info from the DRAM Hole Address Register (section 3.4.8, pub #26094)
495 * for the node represented by mci. Info is passed back in *hole_base,
496 * *hole_offset, and *hole_size. Function returns 0 if info is valid or 1 if
497 * info is invalid. Info may be invalid for either of the following reasons:
498 *
499 * - The revision of the node is not E or greater. In this case, the DRAM Hole
500 * Address Register does not exist.
501 *
502 * - The DramHoleValid bit is cleared in the DRAM Hole Address Register,
503 * indicating that its contents are not valid.
504 *
505 * The values passed back in *hole_base, *hole_offset, and *hole_size are
506 * complete 32-bit values despite the fact that the bitfields in the DHAR
507 * only represent bits 31-24 of the base and offset values.
508 */
509int amd64_get_dram_hole_info(struct mem_ctl_info *mci, u64 *hole_base,
510 u64 *hole_offset, u64 *hole_size)
511{
512 struct amd64_pvt *pvt = mci->pvt_info;
e2ce7255
DT
513
514 /* only revE and later have the DRAM Hole Address Register */
a4b4bedc 515 if (pvt->fam == 0xf && pvt->ext_model < K8_REV_E) {
956b9ba1
JP
516 edac_dbg(1, " revision %d for node %d does not support DHAR\n",
517 pvt->ext_model, pvt->mc_node_id);
e2ce7255
DT
518 return 1;
519 }
520
bc21fa57 521 /* valid for Fam10h and above */
a4b4bedc 522 if (pvt->fam >= 0x10 && !dhar_mem_hoist_valid(pvt)) {
956b9ba1 523 edac_dbg(1, " Dram Memory Hoisting is DISABLED on this system\n");
e2ce7255
DT
524 return 1;
525 }
526
c8e518d5 527 if (!dhar_valid(pvt)) {
956b9ba1
JP
528 edac_dbg(1, " Dram Memory Hoisting is DISABLED on this node %d\n",
529 pvt->mc_node_id);
e2ce7255
DT
530 return 1;
531 }
532
533 /* This node has Memory Hoisting */
534
535 /* +------------------+--------------------+--------------------+-----
536 * | memory | DRAM hole | relocated |
537 * | [0, (x - 1)] | [x, 0xffffffff] | addresses from |
538 * | | | DRAM hole |
539 * | | | [0x100000000, |
540 * | | | (0x100000000+ |
541 * | | | (0xffffffff-x))] |
542 * +------------------+--------------------+--------------------+-----
543 *
544 * Above is a diagram of physical memory showing the DRAM hole and the
545 * relocated addresses from the DRAM hole. As shown, the DRAM hole
546 * starts at address x (the base address) and extends through address
547 * 0xffffffff. The DRAM Hole Address Register (DHAR) relocates the
548 * addresses in the hole so that they start at 0x100000000.
549 */
550
1f31677e
BP
551 *hole_base = dhar_base(pvt);
552 *hole_size = (1ULL << 32) - *hole_base;
e2ce7255 553
a4b4bedc
BP
554 *hole_offset = (pvt->fam > 0xf) ? f10_dhar_offset(pvt)
555 : k8_dhar_offset(pvt);
e2ce7255 556
956b9ba1
JP
557 edac_dbg(1, " DHAR info for node %d base 0x%lx offset 0x%lx size 0x%lx\n",
558 pvt->mc_node_id, (unsigned long)*hole_base,
559 (unsigned long)*hole_offset, (unsigned long)*hole_size);
e2ce7255
DT
560
561 return 0;
562}
563EXPORT_SYMBOL_GPL(amd64_get_dram_hole_info);
564
93c2df58
DT
565/*
566 * Return the DramAddr that the SysAddr given by @sys_addr maps to. It is
567 * assumed that sys_addr maps to the node given by mci.
568 *
569 * The first part of section 3.4.4 (p. 70) shows how the DRAM Base (section
570 * 3.4.4.1) and DRAM Limit (section 3.4.4.2) registers are used to translate a
571 * SysAddr to a DramAddr. If the DRAM Hole Address Register (DHAR) is enabled,
572 * then it is also involved in translating a SysAddr to a DramAddr. Sections
573 * 3.4.8 and 3.5.8.2 describe the DHAR and how it is used for memory hoisting.
574 * These parts of the documentation are unclear. I interpret them as follows:
575 *
576 * When node n receives a SysAddr, it processes the SysAddr as follows:
577 *
578 * 1. It extracts the DRAMBase and DRAMLimit values from the DRAM Base and DRAM
579 * Limit registers for node n. If the SysAddr is not within the range
580 * specified by the base and limit values, then node n ignores the Sysaddr
581 * (since it does not map to node n). Otherwise continue to step 2 below.
582 *
583 * 2. If the DramHoleValid bit of the DHAR for node n is clear, the DHAR is
584 * disabled so skip to step 3 below. Otherwise see if the SysAddr is within
585 * the range of relocated addresses (starting at 0x100000000) from the DRAM
586 * hole. If not, skip to step 3 below. Else get the value of the
587 * DramHoleOffset field from the DHAR. To obtain the DramAddr, subtract the
588 * offset defined by this value from the SysAddr.
589 *
590 * 3. Obtain the base address for node n from the DRAMBase field of the DRAM
591 * Base register for node n. To obtain the DramAddr, subtract the base
592 * address from the SysAddr, as shown near the start of section 3.4.4 (p.70).
593 */
594static u64 sys_addr_to_dram_addr(struct mem_ctl_info *mci, u64 sys_addr)
595{
7f19bf75 596 struct amd64_pvt *pvt = mci->pvt_info;
93c2df58 597 u64 dram_base, hole_base, hole_offset, hole_size, dram_addr;
1f31677e 598 int ret;
93c2df58 599
7f19bf75 600 dram_base = get_dram_base(pvt, pvt->mc_node_id);
93c2df58
DT
601
602 ret = amd64_get_dram_hole_info(mci, &hole_base, &hole_offset,
603 &hole_size);
604 if (!ret) {
1f31677e
BP
605 if ((sys_addr >= (1ULL << 32)) &&
606 (sys_addr < ((1ULL << 32) + hole_size))) {
93c2df58
DT
607 /* use DHAR to translate SysAddr to DramAddr */
608 dram_addr = sys_addr - hole_offset;
609
956b9ba1
JP
610 edac_dbg(2, "using DHAR to translate SysAddr 0x%lx to DramAddr 0x%lx\n",
611 (unsigned long)sys_addr,
612 (unsigned long)dram_addr);
93c2df58
DT
613
614 return dram_addr;
615 }
616 }
617
618 /*
619 * Translate the SysAddr to a DramAddr as shown near the start of
620 * section 3.4.4 (p. 70). Although sys_addr is a 64-bit value, the k8
621 * only deals with 40-bit values. Therefore we discard bits 63-40 of
622 * sys_addr below. If bit 39 of sys_addr is 1 then the bits we
623 * discard are all 1s. Otherwise the bits we discard are all 0s. See
624 * section 3.4.2 of AMD publication 24592: AMD x86-64 Architecture
625 * Programmer's Manual Volume 1 Application Programming.
626 */
10ef6b0d 627 dram_addr = (sys_addr & GENMASK_ULL(39, 0)) - dram_base;
93c2df58 628
956b9ba1
JP
629 edac_dbg(2, "using DRAM Base register to translate SysAddr 0x%lx to DramAddr 0x%lx\n",
630 (unsigned long)sys_addr, (unsigned long)dram_addr);
93c2df58
DT
631 return dram_addr;
632}
633
634/*
635 * @intlv_en is the value of the IntlvEn field from a DRAM Base register
636 * (section 3.4.4.1). Return the number of bits from a SysAddr that are used
637 * for node interleaving.
638 */
639static int num_node_interleave_bits(unsigned intlv_en)
640{
641 static const int intlv_shift_table[] = { 0, 1, 0, 2, 0, 0, 0, 3 };
642 int n;
643
644 BUG_ON(intlv_en > 7);
645 n = intlv_shift_table[intlv_en];
646 return n;
647}
648
649/* Translate the DramAddr given by @dram_addr to an InputAddr. */
650static u64 dram_addr_to_input_addr(struct mem_ctl_info *mci, u64 dram_addr)
651{
652 struct amd64_pvt *pvt;
653 int intlv_shift;
654 u64 input_addr;
655
656 pvt = mci->pvt_info;
657
658 /*
659 * See the start of section 3.4.4 (p. 70, BKDG #26094, K8, revA-E)
660 * concerning translating a DramAddr to an InputAddr.
661 */
7f19bf75 662 intlv_shift = num_node_interleave_bits(dram_intlv_en(pvt, 0));
10ef6b0d 663 input_addr = ((dram_addr >> intlv_shift) & GENMASK_ULL(35, 12)) +
f678b8cc 664 (dram_addr & 0xfff);
93c2df58 665
956b9ba1
JP
666 edac_dbg(2, " Intlv Shift=%d DramAddr=0x%lx maps to InputAddr=0x%lx\n",
667 intlv_shift, (unsigned long)dram_addr,
668 (unsigned long)input_addr);
93c2df58
DT
669
670 return input_addr;
671}
672
673/*
674 * Translate the SysAddr represented by @sys_addr to an InputAddr. It is
675 * assumed that @sys_addr maps to the node given by mci.
676 */
677static u64 sys_addr_to_input_addr(struct mem_ctl_info *mci, u64 sys_addr)
678{
679 u64 input_addr;
680
681 input_addr =
682 dram_addr_to_input_addr(mci, sys_addr_to_dram_addr(mci, sys_addr));
683
c19ca6cb 684 edac_dbg(2, "SysAddr 0x%lx translates to InputAddr 0x%lx\n",
956b9ba1 685 (unsigned long)sys_addr, (unsigned long)input_addr);
93c2df58
DT
686
687 return input_addr;
688}
689
93c2df58
DT
690/* Map the Error address to a PAGE and PAGE OFFSET. */
691static inline void error_address_to_page_and_offset(u64 error_address,
33ca0643 692 struct err_info *err)
93c2df58 693{
33ca0643
BP
694 err->page = (u32) (error_address >> PAGE_SHIFT);
695 err->offset = ((u32) error_address) & ~PAGE_MASK;
93c2df58
DT
696}
697
698/*
699 * @sys_addr is an error address (a SysAddr) extracted from the MCA NB Address
700 * Low (section 3.6.4.5) and MCA NB Address High (section 3.6.4.6) registers
701 * of a node that detected an ECC memory error. mci represents the node that
702 * the error address maps to (possibly different from the node that detected
703 * the error). Return the number of the csrow that sys_addr maps to, or -1 on
704 * error.
705 */
706static int sys_addr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr)
707{
708 int csrow;
709
710 csrow = input_addr_to_csrow(mci, sys_addr_to_input_addr(mci, sys_addr));
711
712 if (csrow == -1)
24f9a7fe
BP
713 amd64_mc_err(mci, "Failed to translate InputAddr to csrow for "
714 "address 0x%lx\n", (unsigned long)sys_addr);
93c2df58
DT
715 return csrow;
716}
e2ce7255 717
bfc04aec 718static int get_channel_from_ecc_syndrome(struct mem_ctl_info *, u16);
2da11654 719
2da11654
DT
720/*
721 * Determine if the DIMMs have ECC enabled. ECC is enabled ONLY if all the DIMMs
722 * are ECC capable.
723 */
d1ea71cd 724static unsigned long determine_edac_cap(struct amd64_pvt *pvt)
2da11654 725{
1f6189ed 726 unsigned long edac_cap = EDAC_FLAG_NONE;
d27f3a34
YG
727 u8 bit;
728
729 if (pvt->umc) {
730 u8 i, umc_en_mask = 0, dimm_ecc_en_mask = 0;
2da11654 731
4d30d2bc 732 for_each_umc(i) {
d27f3a34
YG
733 if (!(pvt->umc[i].sdp_ctrl & UMC_SDP_INIT))
734 continue;
2da11654 735
d27f3a34
YG
736 umc_en_mask |= BIT(i);
737
738 /* UMC Configuration bit 12 (DimmEccEn) */
739 if (pvt->umc[i].umc_cfg & BIT(12))
740 dimm_ecc_en_mask |= BIT(i);
741 }
742
743 if (umc_en_mask == dimm_ecc_en_mask)
744 edac_cap = EDAC_FLAG_SECDED;
745 } else {
746 bit = (pvt->fam > 0xf || pvt->ext_model >= K8_REV_F)
747 ? 19
748 : 17;
749
750 if (pvt->dclr0 & BIT(bit))
751 edac_cap = EDAC_FLAG_SECDED;
752 }
2da11654
DT
753
754 return edac_cap;
755}
756
d1ea71cd 757static void debug_display_dimm_sizes(struct amd64_pvt *, u8);
2da11654 758
d1ea71cd 759static void debug_dump_dramcfg_low(struct amd64_pvt *pvt, u32 dclr, int chan)
68798e17 760{
956b9ba1 761 edac_dbg(1, "F2x%d90 (DRAM Cfg Low): 0x%08x\n", chan, dclr);
68798e17 762
a597d2a5
AG
763 if (pvt->dram_type == MEM_LRDDR3) {
764 u32 dcsm = pvt->csels[chan].csmasks[0];
765 /*
766 * It's assumed all LRDIMMs in a DCT are going to be of
767 * same 'type' until proven otherwise. So, use a cs
768 * value of '0' here to get dcsm value.
769 */
770 edac_dbg(1, " LRDIMM %dx rank multiply\n", (dcsm & 0x3));
771 }
772
773 edac_dbg(1, "All DIMMs support ECC:%s\n",
774 (dclr & BIT(19)) ? "yes" : "no");
775
68798e17 776
956b9ba1
JP
777 edac_dbg(1, " PAR/ERR parity: %s\n",
778 (dclr & BIT(8)) ? "enabled" : "disabled");
68798e17 779
a4b4bedc 780 if (pvt->fam == 0x10)
956b9ba1
JP
781 edac_dbg(1, " DCT 128bit mode width: %s\n",
782 (dclr & BIT(11)) ? "128b" : "64b");
68798e17 783
956b9ba1
JP
784 edac_dbg(1, " x4 logical DIMMs present: L0: %s L1: %s L2: %s L3: %s\n",
785 (dclr & BIT(12)) ? "yes" : "no",
786 (dclr & BIT(13)) ? "yes" : "no",
787 (dclr & BIT(14)) ? "yes" : "no",
788 (dclr & BIT(15)) ? "yes" : "no");
68798e17
BP
789}
790
fc00c6a4
YG
791/*
792 * The Address Mask should be a contiguous set of bits in the non-interleaved
793 * case. So to check for CS interleaving, find the most- and least-significant
794 * bits of the mask, generate a contiguous bitmask, and compare the two.
795 */
796static bool f17_cs_interleaved(struct amd64_pvt *pvt, u8 ctrl, int cs)
797{
798 u32 mask = pvt->csels[ctrl].csmasks[cs >> 1];
799 u32 msb = fls(mask) - 1, lsb = ffs(mask) - 1;
800 u32 test_mask = GENMASK(msb, lsb);
801
802 edac_dbg(1, "mask=0x%08x test_mask=0x%08x\n", mask, test_mask);
803
804 return mask ^ test_mask;
805}
806
07ed82ef
YG
807static void debug_display_dimm_sizes_df(struct amd64_pvt *pvt, u8 ctrl)
808{
eb77e6b8 809 int dimm, size0, size1, cs0, cs1;
07ed82ef
YG
810
811 edac_printk(KERN_DEBUG, EDAC_MC, "UMC%d chip selects:\n", ctrl);
812
05f664ac 813 for (dimm = 0; dimm < 2; dimm++) {
07ed82ef 814 size0 = 0;
eb77e6b8 815 cs0 = dimm * 2;
07ed82ef 816
eb77e6b8
YG
817 if (csrow_enabled(cs0, ctrl, pvt))
818 size0 = pvt->ops->dbam_to_cs(pvt, ctrl, 0, cs0);
07ed82ef
YG
819
820 size1 = 0;
eb77e6b8
YG
821 cs1 = dimm * 2 + 1;
822
fc00c6a4
YG
823 if (csrow_enabled(cs1, ctrl, pvt)) {
824 /*
825 * CS interleaving is only supported if both CSes have
826 * the same amount of memory. Because they are
827 * interleaved, it will look like both CSes have the
828 * full amount of memory. Save the size for both as
829 * half the amount we found on CS0, if interleaved.
830 */
831 if (f17_cs_interleaved(pvt, ctrl, cs1))
832 size1 = size0 = (size0 >> 1);
833 else
834 size1 = pvt->ops->dbam_to_cs(pvt, ctrl, 0, cs1);
835 }
07ed82ef
YG
836
837 amd64_info(EDAC_MC ": %d: %5dMB %d: %5dMB\n",
eb77e6b8
YG
838 cs0, size0,
839 cs1, size1);
07ed82ef
YG
840 }
841}
842
843static void __dump_misc_regs_df(struct amd64_pvt *pvt)
844{
845 struct amd64_umc *umc;
846 u32 i, tmp, umc_base;
847
4d30d2bc 848 for_each_umc(i) {
07ed82ef
YG
849 umc_base = get_umc_base(i);
850 umc = &pvt->umc[i];
851
852 edac_dbg(1, "UMC%d DIMM cfg: 0x%x\n", i, umc->dimm_cfg);
853 edac_dbg(1, "UMC%d UMC cfg: 0x%x\n", i, umc->umc_cfg);
854 edac_dbg(1, "UMC%d SDP ctrl: 0x%x\n", i, umc->sdp_ctrl);
855 edac_dbg(1, "UMC%d ECC ctrl: 0x%x\n", i, umc->ecc_ctrl);
856
857 amd_smn_read(pvt->mc_node_id, umc_base + UMCCH_ECC_BAD_SYMBOL, &tmp);
858 edac_dbg(1, "UMC%d ECC bad symbol: 0x%x\n", i, tmp);
859
860 amd_smn_read(pvt->mc_node_id, umc_base + UMCCH_UMC_CAP, &tmp);
861 edac_dbg(1, "UMC%d UMC cap: 0x%x\n", i, tmp);
862 edac_dbg(1, "UMC%d UMC cap high: 0x%x\n", i, umc->umc_cap_hi);
863
864 edac_dbg(1, "UMC%d ECC capable: %s, ChipKill ECC capable: %s\n",
865 i, (umc->umc_cap_hi & BIT(30)) ? "yes" : "no",
866 (umc->umc_cap_hi & BIT(31)) ? "yes" : "no");
867 edac_dbg(1, "UMC%d All DIMMs support ECC: %s\n",
868 i, (umc->umc_cfg & BIT(12)) ? "yes" : "no");
869 edac_dbg(1, "UMC%d x4 DIMMs present: %s\n",
870 i, (umc->dimm_cfg & BIT(6)) ? "yes" : "no");
871 edac_dbg(1, "UMC%d x16 DIMMs present: %s\n",
872 i, (umc->dimm_cfg & BIT(7)) ? "yes" : "no");
873
874 if (pvt->dram_type == MEM_LRDDR4) {
875 amd_smn_read(pvt->mc_node_id, umc_base + UMCCH_ADDR_CFG, &tmp);
876 edac_dbg(1, "UMC%d LRDIMM %dx rank multiply\n",
877 i, 1 << ((tmp >> 4) & 0x3));
878 }
879
880 debug_display_dimm_sizes_df(pvt, i);
881 }
882
883 edac_dbg(1, "F0x104 (DRAM Hole Address): 0x%08x, base: 0x%08x\n",
884 pvt->dhar, dhar_base(pvt));
885}
886
2da11654 887/* Display and decode various NB registers for debug purposes. */
07ed82ef 888static void __dump_misc_regs(struct amd64_pvt *pvt)
2da11654 889{
956b9ba1 890 edac_dbg(1, "F3xE8 (NB Cap): 0x%08x\n", pvt->nbcap);
68798e17 891
956b9ba1
JP
892 edac_dbg(1, " NB two channel DRAM capable: %s\n",
893 (pvt->nbcap & NBCAP_DCT_DUAL) ? "yes" : "no");
2da11654 894
956b9ba1
JP
895 edac_dbg(1, " ECC capable: %s, ChipKill ECC capable: %s\n",
896 (pvt->nbcap & NBCAP_SECDED) ? "yes" : "no",
897 (pvt->nbcap & NBCAP_CHIPKILL) ? "yes" : "no");
68798e17 898
d1ea71cd 899 debug_dump_dramcfg_low(pvt, pvt->dclr0, 0);
2da11654 900
956b9ba1 901 edac_dbg(1, "F3xB0 (Online Spare): 0x%08x\n", pvt->online_spare);
2da11654 902
956b9ba1
JP
903 edac_dbg(1, "F1xF0 (DRAM Hole Address): 0x%08x, base: 0x%08x, offset: 0x%08x\n",
904 pvt->dhar, dhar_base(pvt),
a4b4bedc
BP
905 (pvt->fam == 0xf) ? k8_dhar_offset(pvt)
906 : f10_dhar_offset(pvt));
2da11654 907
d1ea71cd 908 debug_display_dimm_sizes(pvt, 0);
4d796364 909
8de1d91e 910 /* everything below this point is Fam10h and above */
a4b4bedc 911 if (pvt->fam == 0xf)
2da11654 912 return;
4d796364 913
d1ea71cd 914 debug_display_dimm_sizes(pvt, 1);
2da11654 915
8de1d91e 916 /* Only if NOT ganged does dclr1 have valid info */
68798e17 917 if (!dct_ganging_enabled(pvt))
d1ea71cd 918 debug_dump_dramcfg_low(pvt, pvt->dclr1, 1);
2da11654
DT
919}
920
07ed82ef
YG
921/* Display and decode various NB registers for debug purposes. */
922static void dump_misc_regs(struct amd64_pvt *pvt)
923{
924 if (pvt->umc)
925 __dump_misc_regs_df(pvt);
926 else
927 __dump_misc_regs(pvt);
928
929 edac_dbg(1, " DramHoleValid: %s\n", dhar_valid(pvt) ? "yes" : "no");
930
7835961d 931 amd64_info("using x%u syndromes.\n", pvt->ecc_sym_sz);
07ed82ef
YG
932}
933
94be4bff 934/*
18b94f66 935 * See BKDG, F2x[1,0][5C:40], F2[1,0][6C:60]
94be4bff 936 */
11c75ead 937static void prep_chip_selects(struct amd64_pvt *pvt)
94be4bff 938{
18b94f66 939 if (pvt->fam == 0xf && pvt->ext_model < K8_REV_F) {
11c75ead
BP
940 pvt->csels[0].b_cnt = pvt->csels[1].b_cnt = 8;
941 pvt->csels[0].m_cnt = pvt->csels[1].m_cnt = 8;
a597d2a5 942 } else if (pvt->fam == 0x15 && pvt->model == 0x30) {
18b94f66
AG
943 pvt->csels[0].b_cnt = pvt->csels[1].b_cnt = 4;
944 pvt->csels[0].m_cnt = pvt->csels[1].m_cnt = 2;
05f664ac
YG
945 } else if (pvt->fam >= 0x17) {
946 int umc;
947
948 for_each_umc(umc) {
949 pvt->csels[umc].b_cnt = 4;
950 pvt->csels[umc].m_cnt = 2;
951 }
952
9d858bb1 953 } else {
11c75ead
BP
954 pvt->csels[0].b_cnt = pvt->csels[1].b_cnt = 8;
955 pvt->csels[0].m_cnt = pvt->csels[1].m_cnt = 4;
94be4bff
DT
956 }
957}
958
05f664ac
YG
959static void read_umc_base_mask(struct amd64_pvt *pvt)
960{
961 u32 umc_base_reg, umc_mask_reg;
962 u32 base_reg, mask_reg;
963 u32 *base, *mask;
964 int cs, umc;
965
966 for_each_umc(umc) {
967 umc_base_reg = get_umc_base(umc) + UMCCH_BASE_ADDR;
968
969 for_each_chip_select(cs, umc, pvt) {
970 base = &pvt->csels[umc].csbases[cs];
971
972 base_reg = umc_base_reg + (cs * 4);
973
974 if (!amd_smn_read(pvt->mc_node_id, base_reg, base))
975 edac_dbg(0, " DCSB%d[%d]=0x%08x reg: 0x%x\n",
976 umc, cs, *base, base_reg);
977 }
978
979 umc_mask_reg = get_umc_base(umc) + UMCCH_ADDR_MASK;
980
981 for_each_chip_select_mask(cs, umc, pvt) {
982 mask = &pvt->csels[umc].csmasks[cs];
983
984 mask_reg = umc_mask_reg + (cs * 4);
985
986 if (!amd_smn_read(pvt->mc_node_id, mask_reg, mask))
987 edac_dbg(0, " DCSM%d[%d]=0x%08x reg: 0x%x\n",
988 umc, cs, *mask, mask_reg);
989 }
990 }
991}
992
94be4bff 993/*
11c75ead 994 * Function 2 Offset F10_DCSB0; read in the DCS Base and DCS Mask registers
94be4bff 995 */
b2b0c605 996static void read_dct_base_mask(struct amd64_pvt *pvt)
94be4bff 997{
05f664ac 998 int cs;
94be4bff 999
11c75ead 1000 prep_chip_selects(pvt);
94be4bff 1001
05f664ac
YG
1002 if (pvt->umc)
1003 return read_umc_base_mask(pvt);
b64ce7cd 1004
11c75ead 1005 for_each_chip_select(cs, 0, pvt) {
05f664ac
YG
1006 int reg0 = DCSB0 + (cs * 4);
1007 int reg1 = DCSB1 + (cs * 4);
11c75ead
BP
1008 u32 *base0 = &pvt->csels[0].csbases[cs];
1009 u32 *base1 = &pvt->csels[1].csbases[cs];
b2b0c605 1010
05f664ac
YG
1011 if (!amd64_read_dct_pci_cfg(pvt, 0, reg0, base0))
1012 edac_dbg(0, " DCSB0[%d]=0x%08x reg: F2x%x\n",
1013 cs, *base0, reg0);
8de9930a 1014
05f664ac
YG
1015 if (pvt->fam == 0xf)
1016 continue;
b64ce7cd 1017
05f664ac
YG
1018 if (!amd64_read_dct_pci_cfg(pvt, 1, reg0, base1))
1019 edac_dbg(0, " DCSB1[%d]=0x%08x reg: F2x%x\n",
1020 cs, *base1, (pvt->fam == 0x10) ? reg1
1021 : reg0);
94be4bff
DT
1022 }
1023
11c75ead 1024 for_each_chip_select_mask(cs, 0, pvt) {
05f664ac
YG
1025 int reg0 = DCSM0 + (cs * 4);
1026 int reg1 = DCSM1 + (cs * 4);
11c75ead
BP
1027 u32 *mask0 = &pvt->csels[0].csmasks[cs];
1028 u32 *mask1 = &pvt->csels[1].csmasks[cs];
b2b0c605 1029
05f664ac
YG
1030 if (!amd64_read_dct_pci_cfg(pvt, 0, reg0, mask0))
1031 edac_dbg(0, " DCSM0[%d]=0x%08x reg: F2x%x\n",
1032 cs, *mask0, reg0);
b64ce7cd 1033
05f664ac
YG
1034 if (pvt->fam == 0xf)
1035 continue;
8de9930a 1036
05f664ac
YG
1037 if (!amd64_read_dct_pci_cfg(pvt, 1, reg0, mask1))
1038 edac_dbg(0, " DCSM1[%d]=0x%08x reg: F2x%x\n",
1039 cs, *mask1, (pvt->fam == 0x10) ? reg1
1040 : reg0);
94be4bff
DT
1041 }
1042}
1043
a597d2a5 1044static void determine_memory_type(struct amd64_pvt *pvt)
94be4bff 1045{
a597d2a5 1046 u32 dram_ctrl, dcsm;
94be4bff 1047
a597d2a5
AG
1048 switch (pvt->fam) {
1049 case 0xf:
1050 if (pvt->ext_model >= K8_REV_F)
1051 goto ddr3;
1052
1053 pvt->dram_type = (pvt->dclr0 & BIT(18)) ? MEM_DDR : MEM_RDDR;
1054 return;
1055
1056 case 0x10:
6b4c0bde 1057 if (pvt->dchr0 & DDR3_MODE)
a597d2a5
AG
1058 goto ddr3;
1059
1060 pvt->dram_type = (pvt->dclr0 & BIT(16)) ? MEM_DDR2 : MEM_RDDR2;
1061 return;
1062
1063 case 0x15:
1064 if (pvt->model < 0x60)
1065 goto ddr3;
1066
1067 /*
1068 * Model 0x60h needs special handling:
1069 *
1070 * We use a Chip Select value of '0' to obtain dcsm.
1071 * Theoretically, it is possible to populate LRDIMMs of different
1072 * 'Rank' value on a DCT. But this is not the common case. So,
1073 * it's reasonable to assume all DIMMs are going to be of same
1074 * 'type' until proven otherwise.
1075 */
1076 amd64_read_dct_pci_cfg(pvt, 0, DRAM_CONTROL, &dram_ctrl);
1077 dcsm = pvt->csels[0].csmasks[0];
1078
1079 if (((dram_ctrl >> 8) & 0x7) == 0x2)
1080 pvt->dram_type = MEM_DDR4;
1081 else if (pvt->dclr0 & BIT(16))
1082 pvt->dram_type = MEM_DDR3;
1083 else if (dcsm & 0x3)
1084 pvt->dram_type = MEM_LRDDR3;
6b4c0bde 1085 else
a597d2a5 1086 pvt->dram_type = MEM_RDDR3;
94be4bff 1087
a597d2a5
AG
1088 return;
1089
1090 case 0x16:
1091 goto ddr3;
1092
b64ce7cd 1093 case 0x17:
c4a3e946 1094 case 0x18:
b64ce7cd
YG
1095 if ((pvt->umc[0].dimm_cfg | pvt->umc[1].dimm_cfg) & BIT(5))
1096 pvt->dram_type = MEM_LRDDR4;
1097 else if ((pvt->umc[0].dimm_cfg | pvt->umc[1].dimm_cfg) & BIT(4))
1098 pvt->dram_type = MEM_RDDR4;
1099 else
1100 pvt->dram_type = MEM_DDR4;
1101 return;
1102
a597d2a5
AG
1103 default:
1104 WARN(1, KERN_ERR "%s: Family??? 0x%x\n", __func__, pvt->fam);
1105 pvt->dram_type = MEM_EMPTY;
1106 }
1107 return;
94be4bff 1108
a597d2a5
AG
1109ddr3:
1110 pvt->dram_type = (pvt->dclr0 & BIT(16)) ? MEM_DDR3 : MEM_RDDR3;
94be4bff
DT
1111}
1112
cb328507 1113/* Get the number of DCT channels the memory controller is using. */
ddff876d
DT
1114static int k8_early_channel_count(struct amd64_pvt *pvt)
1115{
cb328507 1116 int flag;
ddff876d 1117
9f56da0e 1118 if (pvt->ext_model >= K8_REV_F)
ddff876d 1119 /* RevF (NPT) and later */
41d8bfab 1120 flag = pvt->dclr0 & WIDTH_128;
9f56da0e 1121 else
ddff876d
DT
1122 /* RevE and earlier */
1123 flag = pvt->dclr0 & REVE_WIDTH_128;
ddff876d
DT
1124
1125 /* not used */
1126 pvt->dclr1 = 0;
1127
1128 return (flag) ? 2 : 1;
1129}
1130
70046624 1131/* On F10h and later ErrAddr is MC4_ADDR[47:1] */
a4b4bedc 1132static u64 get_error_address(struct amd64_pvt *pvt, struct mce *m)
ddff876d 1133{
2ec591ac
BP
1134 u16 mce_nid = amd_get_nb_id(m->extcpu);
1135 struct mem_ctl_info *mci;
70046624
BP
1136 u8 start_bit = 1;
1137 u8 end_bit = 47;
2ec591ac
BP
1138 u64 addr;
1139
1140 mci = edac_mc_find(mce_nid);
1141 if (!mci)
1142 return 0;
1143
1144 pvt = mci->pvt_info;
70046624 1145
a4b4bedc 1146 if (pvt->fam == 0xf) {
70046624
BP
1147 start_bit = 3;
1148 end_bit = 39;
1149 }
1150
10ef6b0d 1151 addr = m->addr & GENMASK_ULL(end_bit, start_bit);
c1ae6830
BP
1152
1153 /*
1154 * Erratum 637 workaround
1155 */
a4b4bedc 1156 if (pvt->fam == 0x15) {
c1ae6830
BP
1157 u64 cc6_base, tmp_addr;
1158 u32 tmp;
8b84c8df 1159 u8 intlv_en;
c1ae6830 1160
10ef6b0d 1161 if ((addr & GENMASK_ULL(47, 24)) >> 24 != 0x00fdf7)
c1ae6830
BP
1162 return addr;
1163
c1ae6830
BP
1164
1165 amd64_read_pci_cfg(pvt->F1, DRAM_LOCAL_NODE_LIM, &tmp);
1166 intlv_en = tmp >> 21 & 0x7;
1167
1168 /* add [47:27] + 3 trailing bits */
10ef6b0d 1169 cc6_base = (tmp & GENMASK_ULL(20, 0)) << 3;
c1ae6830
BP
1170
1171 /* reverse and add DramIntlvEn */
1172 cc6_base |= intlv_en ^ 0x7;
1173
1174 /* pin at [47:24] */
1175 cc6_base <<= 24;
1176
1177 if (!intlv_en)
10ef6b0d 1178 return cc6_base | (addr & GENMASK_ULL(23, 0));
c1ae6830
BP
1179
1180 amd64_read_pci_cfg(pvt->F1, DRAM_LOCAL_NODE_BASE, &tmp);
1181
1182 /* faster log2 */
10ef6b0d 1183 tmp_addr = (addr & GENMASK_ULL(23, 12)) << __fls(intlv_en + 1);
c1ae6830
BP
1184
1185 /* OR DramIntlvSel into bits [14:12] */
10ef6b0d 1186 tmp_addr |= (tmp & GENMASK_ULL(23, 21)) >> 9;
c1ae6830
BP
1187
1188 /* add remaining [11:0] bits from original MC4_ADDR */
10ef6b0d 1189 tmp_addr |= addr & GENMASK_ULL(11, 0);
c1ae6830
BP
1190
1191 return cc6_base | tmp_addr;
1192 }
1193
1194 return addr;
ddff876d
DT
1195}
1196
e2c0bffe
DB
1197static struct pci_dev *pci_get_related_function(unsigned int vendor,
1198 unsigned int device,
1199 struct pci_dev *related)
1200{
1201 struct pci_dev *dev = NULL;
1202
1203 while ((dev = pci_get_device(vendor, device, dev))) {
1204 if (pci_domain_nr(dev->bus) == pci_domain_nr(related->bus) &&
1205 (dev->bus->number == related->bus->number) &&
1206 (PCI_SLOT(dev->devfn) == PCI_SLOT(related->devfn)))
1207 break;
1208 }
1209
1210 return dev;
1211}
1212
7f19bf75 1213static void read_dram_base_limit_regs(struct amd64_pvt *pvt, unsigned range)
ddff876d 1214{
e2c0bffe 1215 struct amd_northbridge *nb;
18b94f66
AG
1216 struct pci_dev *f1 = NULL;
1217 unsigned int pci_func;
71d2a32e 1218 int off = range << 3;
e2c0bffe 1219 u32 llim;
ddff876d 1220
7f19bf75
BP
1221 amd64_read_pci_cfg(pvt->F1, DRAM_BASE_LO + off, &pvt->ranges[range].base.lo);
1222 amd64_read_pci_cfg(pvt->F1, DRAM_LIMIT_LO + off, &pvt->ranges[range].lim.lo);
ddff876d 1223
18b94f66 1224 if (pvt->fam == 0xf)
7f19bf75 1225 return;
ddff876d 1226
7f19bf75
BP
1227 if (!dram_rw(pvt, range))
1228 return;
ddff876d 1229
7f19bf75
BP
1230 amd64_read_pci_cfg(pvt->F1, DRAM_BASE_HI + off, &pvt->ranges[range].base.hi);
1231 amd64_read_pci_cfg(pvt->F1, DRAM_LIMIT_HI + off, &pvt->ranges[range].lim.hi);
f08e457c 1232
e2c0bffe 1233 /* F15h: factor in CC6 save area by reading dst node's limit reg */
18b94f66 1234 if (pvt->fam != 0x15)
e2c0bffe 1235 return;
f08e457c 1236
e2c0bffe
DB
1237 nb = node_to_amd_nb(dram_dst_node(pvt, range));
1238 if (WARN_ON(!nb))
1239 return;
f08e457c 1240
a597d2a5
AG
1241 if (pvt->model == 0x60)
1242 pci_func = PCI_DEVICE_ID_AMD_15H_M60H_NB_F1;
1243 else if (pvt->model == 0x30)
1244 pci_func = PCI_DEVICE_ID_AMD_15H_M30H_NB_F1;
1245 else
1246 pci_func = PCI_DEVICE_ID_AMD_15H_NB_F1;
18b94f66
AG
1247
1248 f1 = pci_get_related_function(nb->misc->vendor, pci_func, nb->misc);
e2c0bffe
DB
1249 if (WARN_ON(!f1))
1250 return;
f08e457c 1251
e2c0bffe 1252 amd64_read_pci_cfg(f1, DRAM_LOCAL_NODE_LIM, &llim);
f08e457c 1253
10ef6b0d 1254 pvt->ranges[range].lim.lo &= GENMASK_ULL(15, 0);
f08e457c 1255
e2c0bffe
DB
1256 /* {[39:27],111b} */
1257 pvt->ranges[range].lim.lo |= ((llim & 0x1fff) << 3 | 0x7) << 16;
f08e457c 1258
10ef6b0d 1259 pvt->ranges[range].lim.hi &= GENMASK_ULL(7, 0);
f08e457c 1260
e2c0bffe
DB
1261 /* [47:40] */
1262 pvt->ranges[range].lim.hi |= llim >> 13;
1263
1264 pci_dev_put(f1);
ddff876d
DT
1265}
1266
f192c7b1 1267static void k8_map_sysaddr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr,
33ca0643 1268 struct err_info *err)
ddff876d 1269{
f192c7b1 1270 struct amd64_pvt *pvt = mci->pvt_info;
ddff876d 1271
33ca0643 1272 error_address_to_page_and_offset(sys_addr, err);
ab5a503c
MCC
1273
1274 /*
1275 * Find out which node the error address belongs to. This may be
1276 * different from the node that detected the error.
1277 */
33ca0643
BP
1278 err->src_mci = find_mc_by_sys_addr(mci, sys_addr);
1279 if (!err->src_mci) {
ab5a503c
MCC
1280 amd64_mc_err(mci, "failed to map error addr 0x%lx to a node\n",
1281 (unsigned long)sys_addr);
33ca0643 1282 err->err_code = ERR_NODE;
ab5a503c
MCC
1283 return;
1284 }
1285
1286 /* Now map the sys_addr to a CSROW */
33ca0643
BP
1287 err->csrow = sys_addr_to_csrow(err->src_mci, sys_addr);
1288 if (err->csrow < 0) {
1289 err->err_code = ERR_CSROW;
ab5a503c
MCC
1290 return;
1291 }
1292
ddff876d 1293 /* CHIPKILL enabled */
f192c7b1 1294 if (pvt->nbcfg & NBCFG_CHIPKILL) {
33ca0643
BP
1295 err->channel = get_channel_from_ecc_syndrome(mci, err->syndrome);
1296 if (err->channel < 0) {
ddff876d
DT
1297 /*
1298 * Syndrome didn't map, so we don't know which of the
1299 * 2 DIMMs is in error. So we need to ID 'both' of them
1300 * as suspect.
1301 */
33ca0643 1302 amd64_mc_warn(err->src_mci, "unknown syndrome 0x%04x - "
ab5a503c 1303 "possible error reporting race\n",
33ca0643
BP
1304 err->syndrome);
1305 err->err_code = ERR_CHANNEL;
ddff876d
DT
1306 return;
1307 }
1308 } else {
1309 /*
1310 * non-chipkill ecc mode
1311 *
1312 * The k8 documentation is unclear about how to determine the
1313 * channel number when using non-chipkill memory. This method
1314 * was obtained from email communication with someone at AMD.
1315 * (Wish the email was placed in this comment - norsk)
1316 */
33ca0643 1317 err->channel = ((sys_addr & BIT(3)) != 0);
ddff876d 1318 }
ddff876d
DT
1319}
1320
41d8bfab 1321static int ddr2_cs_size(unsigned i, bool dct_width)
ddff876d 1322{
41d8bfab 1323 unsigned shift = 0;
ddff876d 1324
41d8bfab
BP
1325 if (i <= 2)
1326 shift = i;
1327 else if (!(i & 0x1))
1328 shift = i >> 1;
1433eb99 1329 else
41d8bfab 1330 shift = (i + 1) >> 1;
ddff876d 1331
41d8bfab
BP
1332 return 128 << (shift + !!dct_width);
1333}
1334
1335static int k8_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
a597d2a5 1336 unsigned cs_mode, int cs_mask_nr)
41d8bfab
BP
1337{
1338 u32 dclr = dct ? pvt->dclr1 : pvt->dclr0;
1339
1340 if (pvt->ext_model >= K8_REV_F) {
1341 WARN_ON(cs_mode > 11);
1342 return ddr2_cs_size(cs_mode, dclr & WIDTH_128);
1343 }
1344 else if (pvt->ext_model >= K8_REV_D) {
11b0a314 1345 unsigned diff;
41d8bfab
BP
1346 WARN_ON(cs_mode > 10);
1347
11b0a314
BP
1348 /*
1349 * the below calculation, besides trying to win an obfuscated C
1350 * contest, maps cs_mode values to DIMM chip select sizes. The
1351 * mappings are:
1352 *
1353 * cs_mode CS size (mb)
1354 * ======= ============
1355 * 0 32
1356 * 1 64
1357 * 2 128
1358 * 3 128
1359 * 4 256
1360 * 5 512
1361 * 6 256
1362 * 7 512
1363 * 8 1024
1364 * 9 1024
1365 * 10 2048
1366 *
1367 * Basically, it calculates a value with which to shift the
1368 * smallest CS size of 32MB.
1369 *
1370 * ddr[23]_cs_size have a similar purpose.
1371 */
1372 diff = cs_mode/3 + (unsigned)(cs_mode > 5);
1373
1374 return 32 << (cs_mode - diff);
41d8bfab
BP
1375 }
1376 else {
1377 WARN_ON(cs_mode > 6);
1378 return 32 << cs_mode;
1379 }
ddff876d
DT
1380}
1381
1afd3c98
DT
1382/*
1383 * Get the number of DCT channels in use.
1384 *
1385 * Return:
1386 * number of Memory Channels in operation
1387 * Pass back:
1388 * contents of the DCL0_LOW register
1389 */
7d20d14d 1390static int f1x_early_channel_count(struct amd64_pvt *pvt)
1afd3c98 1391{
6ba5dcdc 1392 int i, j, channels = 0;
1afd3c98 1393
7d20d14d 1394 /* On F10h, if we are in 128 bit mode, then we are using 2 channels */
a4b4bedc 1395 if (pvt->fam == 0x10 && (pvt->dclr0 & WIDTH_128))
7d20d14d 1396 return 2;
1afd3c98
DT
1397
1398 /*
d16149e8
BP
1399 * Need to check if in unganged mode: In such, there are 2 channels,
1400 * but they are not in 128 bit mode and thus the above 'dclr0' status
1401 * bit will be OFF.
1afd3c98
DT
1402 *
1403 * Need to check DCT0[0] and DCT1[0] to see if only one of them has
1404 * their CSEnable bit on. If so, then SINGLE DIMM case.
1405 */
956b9ba1 1406 edac_dbg(0, "Data width is not 128 bits - need more decoding\n");
ddff876d 1407
1afd3c98
DT
1408 /*
1409 * Check DRAM Bank Address Mapping values for each DIMM to see if there
1410 * is more than just one DIMM present in unganged mode. Need to check
1411 * both controllers since DIMMs can be placed in either one.
1412 */
525a1b20
BP
1413 for (i = 0; i < 2; i++) {
1414 u32 dbam = (i ? pvt->dbam1 : pvt->dbam0);
1afd3c98 1415
57a30854
WW
1416 for (j = 0; j < 4; j++) {
1417 if (DBAM_DIMM(j, dbam) > 0) {
1418 channels++;
1419 break;
1420 }
1421 }
1afd3c98
DT
1422 }
1423
d16149e8
BP
1424 if (channels > 2)
1425 channels = 2;
1426
24f9a7fe 1427 amd64_info("MCT channel count: %d\n", channels);
1afd3c98
DT
1428
1429 return channels;
1afd3c98
DT
1430}
1431
f1cbbec9
YG
1432static int f17_early_channel_count(struct amd64_pvt *pvt)
1433{
1434 int i, channels = 0;
1435
1436 /* SDP Control bit 31 (SdpInit) is clear for unused UMC channels */
4d30d2bc 1437 for_each_umc(i)
f1cbbec9
YG
1438 channels += !!(pvt->umc[i].sdp_ctrl & UMC_SDP_INIT);
1439
1440 amd64_info("MCT channel count: %d\n", channels);
1441
1442 return channels;
1443}
1444
41d8bfab 1445static int ddr3_cs_size(unsigned i, bool dct_width)
1afd3c98 1446{
41d8bfab
BP
1447 unsigned shift = 0;
1448 int cs_size = 0;
1449
1450 if (i == 0 || i == 3 || i == 4)
1451 cs_size = -1;
1452 else if (i <= 2)
1453 shift = i;
1454 else if (i == 12)
1455 shift = 7;
1456 else if (!(i & 0x1))
1457 shift = i >> 1;
1458 else
1459 shift = (i + 1) >> 1;
1460
1461 if (cs_size != -1)
1462 cs_size = (128 * (1 << !!dct_width)) << shift;
1463
1464 return cs_size;
1465}
1466
a597d2a5
AG
1467static int ddr3_lrdimm_cs_size(unsigned i, unsigned rank_multiply)
1468{
1469 unsigned shift = 0;
1470 int cs_size = 0;
1471
1472 if (i < 4 || i == 6)
1473 cs_size = -1;
1474 else if (i == 12)
1475 shift = 7;
1476 else if (!(i & 0x1))
1477 shift = i >> 1;
1478 else
1479 shift = (i + 1) >> 1;
1480
1481 if (cs_size != -1)
1482 cs_size = rank_multiply * (128 << shift);
1483
1484 return cs_size;
1485}
1486
1487static int ddr4_cs_size(unsigned i)
1488{
1489 int cs_size = 0;
1490
1491 if (i == 0)
1492 cs_size = -1;
1493 else if (i == 1)
1494 cs_size = 1024;
1495 else
1496 /* Min cs_size = 1G */
1497 cs_size = 1024 * (1 << (i >> 1));
1498
1499 return cs_size;
1500}
1501
41d8bfab 1502static int f10_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
a597d2a5 1503 unsigned cs_mode, int cs_mask_nr)
41d8bfab
BP
1504{
1505 u32 dclr = dct ? pvt->dclr1 : pvt->dclr0;
1506
1507 WARN_ON(cs_mode > 11);
1433eb99
BP
1508
1509 if (pvt->dchr0 & DDR3_MODE || pvt->dchr1 & DDR3_MODE)
41d8bfab 1510 return ddr3_cs_size(cs_mode, dclr & WIDTH_128);
1433eb99 1511 else
41d8bfab
BP
1512 return ddr2_cs_size(cs_mode, dclr & WIDTH_128);
1513}
1514
1515/*
1516 * F15h supports only 64bit DCT interfaces
1517 */
1518static int f15_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
a597d2a5 1519 unsigned cs_mode, int cs_mask_nr)
41d8bfab
BP
1520{
1521 WARN_ON(cs_mode > 12);
1433eb99 1522
41d8bfab 1523 return ddr3_cs_size(cs_mode, false);
1afd3c98
DT
1524}
1525
a597d2a5
AG
1526/* F15h M60h supports DDR4 mapping as well.. */
1527static int f15_m60h_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
1528 unsigned cs_mode, int cs_mask_nr)
1529{
1530 int cs_size;
1531 u32 dcsm = pvt->csels[dct].csmasks[cs_mask_nr];
1532
1533 WARN_ON(cs_mode > 12);
1534
1535 if (pvt->dram_type == MEM_DDR4) {
1536 if (cs_mode > 9)
1537 return -1;
1538
1539 cs_size = ddr4_cs_size(cs_mode);
1540 } else if (pvt->dram_type == MEM_LRDDR3) {
1541 unsigned rank_multiply = dcsm & 0xf;
1542
1543 if (rank_multiply == 3)
1544 rank_multiply = 4;
1545 cs_size = ddr3_lrdimm_cs_size(cs_mode, rank_multiply);
1546 } else {
1547 /* Minimum cs size is 512mb for F15hM60h*/
1548 if (cs_mode == 0x1)
1549 return -1;
1550
1551 cs_size = ddr3_cs_size(cs_mode, false);
1552 }
1553
1554 return cs_size;
1555}
1556
94c1acf2 1557/*
18b94f66 1558 * F16h and F15h model 30h have only limited cs_modes.
94c1acf2
AG
1559 */
1560static int f16_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
a597d2a5 1561 unsigned cs_mode, int cs_mask_nr)
94c1acf2
AG
1562{
1563 WARN_ON(cs_mode > 12);
1564
1565 if (cs_mode == 6 || cs_mode == 8 ||
1566 cs_mode == 9 || cs_mode == 12)
1567 return -1;
1568 else
1569 return ddr3_cs_size(cs_mode, false);
1570}
1571
f1cbbec9
YG
1572static int f17_base_addr_to_cs_size(struct amd64_pvt *pvt, u8 umc,
1573 unsigned int cs_mode, int csrow_nr)
1574{
1575 u32 base_addr = pvt->csels[umc].csbases[csrow_nr];
1576
1577 /* Each mask is used for every two base addresses. */
1578 u32 addr_mask = pvt->csels[umc].csmasks[csrow_nr >> 1];
1579
1580 /* Register [31:1] = Address [39:9]. Size is in kBs here. */
1581 u32 size = ((addr_mask >> 1) - (base_addr >> 1) + 1) >> 1;
1582
1583 edac_dbg(1, "BaseAddr: 0x%x, AddrMask: 0x%x\n", base_addr, addr_mask);
1584
1585 /* Return size in MBs. */
1586 return size >> 10;
1587}
1588
5a5d2371 1589static void read_dram_ctl_register(struct amd64_pvt *pvt)
6163b5d4 1590{
6163b5d4 1591
a4b4bedc 1592 if (pvt->fam == 0xf)
5a5d2371
BP
1593 return;
1594
7981a28f 1595 if (!amd64_read_pci_cfg(pvt->F2, DCT_SEL_LO, &pvt->dct_sel_lo)) {
956b9ba1
JP
1596 edac_dbg(0, "F2x110 (DCTSelLow): 0x%08x, High range addrs at: 0x%x\n",
1597 pvt->dct_sel_lo, dct_sel_baseaddr(pvt));
72381bd5 1598
956b9ba1
JP
1599 edac_dbg(0, " DCTs operate in %s mode\n",
1600 (dct_ganging_enabled(pvt) ? "ganged" : "unganged"));
72381bd5
BP
1601
1602 if (!dct_ganging_enabled(pvt))
956b9ba1
JP
1603 edac_dbg(0, " Address range split per DCT: %s\n",
1604 (dct_high_range_enabled(pvt) ? "yes" : "no"));
72381bd5 1605
956b9ba1
JP
1606 edac_dbg(0, " data interleave for ECC: %s, DRAM cleared since last warm reset: %s\n",
1607 (dct_data_intlv_enabled(pvt) ? "enabled" : "disabled"),
1608 (dct_memory_cleared(pvt) ? "yes" : "no"));
72381bd5 1609
956b9ba1
JP
1610 edac_dbg(0, " channel interleave: %s, "
1611 "interleave bits selector: 0x%x\n",
1612 (dct_interleave_enabled(pvt) ? "enabled" : "disabled"),
1613 dct_sel_interleave_addr(pvt));
6163b5d4
DT
1614 }
1615
7981a28f 1616 amd64_read_pci_cfg(pvt->F2, DCT_SEL_HI, &pvt->dct_sel_hi);
6163b5d4
DT
1617}
1618
18b94f66
AG
1619/*
1620 * Determine channel (DCT) based on the interleaving mode (see F15h M30h BKDG,
1621 * 2.10.12 Memory Interleaving Modes).
1622 */
1623static u8 f15_m30h_determine_channel(struct amd64_pvt *pvt, u64 sys_addr,
1624 u8 intlv_en, int num_dcts_intlv,
1625 u32 dct_sel)
1626{
1627 u8 channel = 0;
1628 u8 select;
1629
1630 if (!(intlv_en))
1631 return (u8)(dct_sel);
1632
1633 if (num_dcts_intlv == 2) {
1634 select = (sys_addr >> 8) & 0x3;
1635 channel = select ? 0x3 : 0;
9d0e8d83
AG
1636 } else if (num_dcts_intlv == 4) {
1637 u8 intlv_addr = dct_sel_interleave_addr(pvt);
1638 switch (intlv_addr) {
1639 case 0x4:
1640 channel = (sys_addr >> 8) & 0x3;
1641 break;
1642 case 0x5:
1643 channel = (sys_addr >> 9) & 0x3;
1644 break;
1645 }
1646 }
18b94f66
AG
1647 return channel;
1648}
1649
f71d0a05 1650/*
229a7a11 1651 * Determine channel (DCT) based on the interleaving mode: F10h BKDG, 2.8.9 Memory
f71d0a05
DT
1652 * Interleaving Modes.
1653 */
b15f0fca 1654static u8 f1x_determine_channel(struct amd64_pvt *pvt, u64 sys_addr,
229a7a11 1655 bool hi_range_sel, u8 intlv_en)
6163b5d4 1656{
151fa71c 1657 u8 dct_sel_high = (pvt->dct_sel_lo >> 1) & 1;
6163b5d4
DT
1658
1659 if (dct_ganging_enabled(pvt))
229a7a11 1660 return 0;
6163b5d4 1661
229a7a11
BP
1662 if (hi_range_sel)
1663 return dct_sel_high;
6163b5d4 1664
229a7a11
BP
1665 /*
1666 * see F2x110[DctSelIntLvAddr] - channel interleave mode
1667 */
1668 if (dct_interleave_enabled(pvt)) {
1669 u8 intlv_addr = dct_sel_interleave_addr(pvt);
1670
1671 /* return DCT select function: 0=DCT0, 1=DCT1 */
1672 if (!intlv_addr)
1673 return sys_addr >> 6 & 1;
1674
1675 if (intlv_addr & 0x2) {
1676 u8 shift = intlv_addr & 0x1 ? 9 : 6;
dc0a50a8 1677 u32 temp = hweight_long((u32) ((sys_addr >> 16) & 0x1F)) & 1;
229a7a11
BP
1678
1679 return ((sys_addr >> shift) & 1) ^ temp;
1680 }
1681
dc0a50a8
YG
1682 if (intlv_addr & 0x4) {
1683 u8 shift = intlv_addr & 0x1 ? 9 : 8;
1684
1685 return (sys_addr >> shift) & 1;
1686 }
1687
229a7a11
BP
1688 return (sys_addr >> (12 + hweight8(intlv_en))) & 1;
1689 }
1690
1691 if (dct_high_range_enabled(pvt))
1692 return ~dct_sel_high & 1;
6163b5d4
DT
1693
1694 return 0;
1695}
1696
c8e518d5 1697/* Convert the sys_addr to the normalized DCT address */
c7e5301a 1698static u64 f1x_get_norm_dct_addr(struct amd64_pvt *pvt, u8 range,
c8e518d5
BP
1699 u64 sys_addr, bool hi_rng,
1700 u32 dct_sel_base_addr)
6163b5d4
DT
1701{
1702 u64 chan_off;
c8e518d5
BP
1703 u64 dram_base = get_dram_base(pvt, range);
1704 u64 hole_off = f10_dhar_offset(pvt);
6f3508f6 1705 u64 dct_sel_base_off = (u64)(pvt->dct_sel_hi & 0xFFFFFC00) << 16;
6163b5d4 1706
c8e518d5
BP
1707 if (hi_rng) {
1708 /*
1709 * if
1710 * base address of high range is below 4Gb
1711 * (bits [47:27] at [31:11])
1712 * DRAM address space on this DCT is hoisted above 4Gb &&
1713 * sys_addr > 4Gb
1714 *
1715 * remove hole offset from sys_addr
1716 * else
1717 * remove high range offset from sys_addr
1718 */
1719 if ((!(dct_sel_base_addr >> 16) ||
1720 dct_sel_base_addr < dhar_base(pvt)) &&
972ea17a 1721 dhar_valid(pvt) &&
c8e518d5 1722 (sys_addr >= BIT_64(32)))
bc21fa57 1723 chan_off = hole_off;
6163b5d4
DT
1724 else
1725 chan_off = dct_sel_base_off;
1726 } else {
c8e518d5
BP
1727 /*
1728 * if
1729 * we have a valid hole &&
1730 * sys_addr > 4Gb
1731 *
1732 * remove hole
1733 * else
1734 * remove dram base to normalize to DCT address
1735 */
972ea17a 1736 if (dhar_valid(pvt) && (sys_addr >= BIT_64(32)))
bc21fa57 1737 chan_off = hole_off;
6163b5d4 1738 else
c8e518d5 1739 chan_off = dram_base;
6163b5d4
DT
1740 }
1741
10ef6b0d 1742 return (sys_addr & GENMASK_ULL(47,6)) - (chan_off & GENMASK_ULL(47,23));
6163b5d4
DT
1743}
1744
6163b5d4
DT
1745/*
1746 * checks if the csrow passed in is marked as SPARED, if so returns the new
1747 * spare row
1748 */
11c75ead 1749static int f10_process_possible_spare(struct amd64_pvt *pvt, u8 dct, int csrow)
6163b5d4 1750{
614ec9d8
BP
1751 int tmp_cs;
1752
1753 if (online_spare_swap_done(pvt, dct) &&
1754 csrow == online_spare_bad_dramcs(pvt, dct)) {
1755
1756 for_each_chip_select(tmp_cs, dct, pvt) {
1757 if (chip_select_base(tmp_cs, dct, pvt) & 0x2) {
1758 csrow = tmp_cs;
1759 break;
1760 }
1761 }
6163b5d4
DT
1762 }
1763 return csrow;
1764}
1765
1766/*
1767 * Iterate over the DRAM DCT "base" and "mask" registers looking for a
1768 * SystemAddr match on the specified 'ChannelSelect' and 'NodeID'
1769 *
1770 * Return:
1771 * -EINVAL: NOT FOUND
1772 * 0..csrow = Chip-Select Row
1773 */
c7e5301a 1774static int f1x_lookup_addr_in_dct(u64 in_addr, u8 nid, u8 dct)
6163b5d4
DT
1775{
1776 struct mem_ctl_info *mci;
1777 struct amd64_pvt *pvt;
11c75ead 1778 u64 cs_base, cs_mask;
6163b5d4
DT
1779 int cs_found = -EINVAL;
1780 int csrow;
1781
2ec591ac 1782 mci = edac_mc_find(nid);
6163b5d4
DT
1783 if (!mci)
1784 return cs_found;
1785
1786 pvt = mci->pvt_info;
1787
956b9ba1 1788 edac_dbg(1, "input addr: 0x%llx, DCT: %d\n", in_addr, dct);
6163b5d4 1789
11c75ead
BP
1790 for_each_chip_select(csrow, dct, pvt) {
1791 if (!csrow_enabled(csrow, dct, pvt))
6163b5d4
DT
1792 continue;
1793
11c75ead 1794 get_cs_base_and_mask(pvt, csrow, dct, &cs_base, &cs_mask);
6163b5d4 1795
956b9ba1
JP
1796 edac_dbg(1, " CSROW=%d CSBase=0x%llx CSMask=0x%llx\n",
1797 csrow, cs_base, cs_mask);
6163b5d4 1798
11c75ead 1799 cs_mask = ~cs_mask;
6163b5d4 1800
956b9ba1
JP
1801 edac_dbg(1, " (InputAddr & ~CSMask)=0x%llx (CSBase & ~CSMask)=0x%llx\n",
1802 (in_addr & cs_mask), (cs_base & cs_mask));
6163b5d4 1803
11c75ead 1804 if ((in_addr & cs_mask) == (cs_base & cs_mask)) {
18b94f66
AG
1805 if (pvt->fam == 0x15 && pvt->model >= 0x30) {
1806 cs_found = csrow;
1807 break;
1808 }
11c75ead 1809 cs_found = f10_process_possible_spare(pvt, dct, csrow);
6163b5d4 1810
956b9ba1 1811 edac_dbg(1, " MATCH csrow=%d\n", cs_found);
6163b5d4
DT
1812 break;
1813 }
1814 }
1815 return cs_found;
1816}
1817
95b0ef55
BP
1818/*
1819 * See F2x10C. Non-interleaved graphics framebuffer memory under the 16G is
1820 * swapped with a region located at the bottom of memory so that the GPU can use
1821 * the interleaved region and thus two channels.
1822 */
b15f0fca 1823static u64 f1x_swap_interleaved_region(struct amd64_pvt *pvt, u64 sys_addr)
95b0ef55
BP
1824{
1825 u32 swap_reg, swap_base, swap_limit, rgn_size, tmp_addr;
1826
a4b4bedc 1827 if (pvt->fam == 0x10) {
95b0ef55 1828 /* only revC3 and revE have that feature */
a4b4bedc 1829 if (pvt->model < 4 || (pvt->model < 0xa && pvt->stepping < 3))
95b0ef55
BP
1830 return sys_addr;
1831 }
1832
7981a28f 1833 amd64_read_pci_cfg(pvt->F2, SWAP_INTLV_REG, &swap_reg);
95b0ef55
BP
1834
1835 if (!(swap_reg & 0x1))
1836 return sys_addr;
1837
1838 swap_base = (swap_reg >> 3) & 0x7f;
1839 swap_limit = (swap_reg >> 11) & 0x7f;
1840 rgn_size = (swap_reg >> 20) & 0x7f;
1841 tmp_addr = sys_addr >> 27;
1842
1843 if (!(sys_addr >> 34) &&
1844 (((tmp_addr >= swap_base) &&
1845 (tmp_addr <= swap_limit)) ||
1846 (tmp_addr < rgn_size)))
1847 return sys_addr ^ (u64)swap_base << 27;
1848
1849 return sys_addr;
1850}
1851
f71d0a05 1852/* For a given @dram_range, check if @sys_addr falls within it. */
e761359a 1853static int f1x_match_to_this_node(struct amd64_pvt *pvt, unsigned range,
33ca0643 1854 u64 sys_addr, int *chan_sel)
f71d0a05 1855{
229a7a11 1856 int cs_found = -EINVAL;
c8e518d5 1857 u64 chan_addr;
5d4b58e8 1858 u32 dct_sel_base;
11c75ead 1859 u8 channel;
229a7a11 1860 bool high_range = false;
f71d0a05 1861
7f19bf75 1862 u8 node_id = dram_dst_node(pvt, range);
229a7a11 1863 u8 intlv_en = dram_intlv_en(pvt, range);
7f19bf75 1864 u32 intlv_sel = dram_intlv_sel(pvt, range);
f71d0a05 1865
956b9ba1
JP
1866 edac_dbg(1, "(range %d) SystemAddr= 0x%llx Limit=0x%llx\n",
1867 range, sys_addr, get_dram_limit(pvt, range));
f71d0a05 1868
355fba60
BP
1869 if (dhar_valid(pvt) &&
1870 dhar_base(pvt) <= sys_addr &&
1871 sys_addr < BIT_64(32)) {
1872 amd64_warn("Huh? Address is in the MMIO hole: 0x%016llx\n",
1873 sys_addr);
1874 return -EINVAL;
1875 }
1876
f030ddfb 1877 if (intlv_en && (intlv_sel != ((sys_addr >> 12) & intlv_en)))
f71d0a05
DT
1878 return -EINVAL;
1879
b15f0fca 1880 sys_addr = f1x_swap_interleaved_region(pvt, sys_addr);
95b0ef55 1881
f71d0a05
DT
1882 dct_sel_base = dct_sel_baseaddr(pvt);
1883
1884 /*
1885 * check whether addresses >= DctSelBaseAddr[47:27] are to be used to
1886 * select between DCT0 and DCT1.
1887 */
1888 if (dct_high_range_enabled(pvt) &&
1889 !dct_ganging_enabled(pvt) &&
1890 ((sys_addr >> 27) >= (dct_sel_base >> 11)))
229a7a11 1891 high_range = true;
f71d0a05 1892
b15f0fca 1893 channel = f1x_determine_channel(pvt, sys_addr, high_range, intlv_en);
f71d0a05 1894
b15f0fca 1895 chan_addr = f1x_get_norm_dct_addr(pvt, range, sys_addr,
c8e518d5 1896 high_range, dct_sel_base);
f71d0a05 1897
e2f79dbd
BP
1898 /* Remove node interleaving, see F1x120 */
1899 if (intlv_en)
1900 chan_addr = ((chan_addr >> (12 + hweight8(intlv_en))) << 12) |
1901 (chan_addr & 0xfff);
f71d0a05 1902
5d4b58e8 1903 /* remove channel interleave */
f71d0a05
DT
1904 if (dct_interleave_enabled(pvt) &&
1905 !dct_high_range_enabled(pvt) &&
1906 !dct_ganging_enabled(pvt)) {
5d4b58e8
BP
1907
1908 if (dct_sel_interleave_addr(pvt) != 1) {
1909 if (dct_sel_interleave_addr(pvt) == 0x3)
1910 /* hash 9 */
1911 chan_addr = ((chan_addr >> 10) << 9) |
1912 (chan_addr & 0x1ff);
1913 else
1914 /* A[6] or hash 6 */
1915 chan_addr = ((chan_addr >> 7) << 6) |
1916 (chan_addr & 0x3f);
1917 } else
1918 /* A[12] */
1919 chan_addr = ((chan_addr >> 13) << 12) |
1920 (chan_addr & 0xfff);
f71d0a05
DT
1921 }
1922
956b9ba1 1923 edac_dbg(1, " Normalized DCT addr: 0x%llx\n", chan_addr);
f71d0a05 1924
b15f0fca 1925 cs_found = f1x_lookup_addr_in_dct(chan_addr, node_id, channel);
f71d0a05 1926
33ca0643 1927 if (cs_found >= 0)
f71d0a05 1928 *chan_sel = channel;
33ca0643 1929
f71d0a05
DT
1930 return cs_found;
1931}
1932
18b94f66
AG
1933static int f15_m30h_match_to_this_node(struct amd64_pvt *pvt, unsigned range,
1934 u64 sys_addr, int *chan_sel)
1935{
1936 int cs_found = -EINVAL;
1937 int num_dcts_intlv = 0;
1938 u64 chan_addr, chan_offset;
1939 u64 dct_base, dct_limit;
1940 u32 dct_cont_base_reg, dct_cont_limit_reg, tmp;
1941 u8 channel, alias_channel, leg_mmio_hole, dct_sel, dct_offset_en;
1942
1943 u64 dhar_offset = f10_dhar_offset(pvt);
1944 u8 intlv_addr = dct_sel_interleave_addr(pvt);
1945 u8 node_id = dram_dst_node(pvt, range);
1946 u8 intlv_en = dram_intlv_en(pvt, range);
1947
1948 amd64_read_pci_cfg(pvt->F1, DRAM_CONT_BASE, &dct_cont_base_reg);
1949 amd64_read_pci_cfg(pvt->F1, DRAM_CONT_LIMIT, &dct_cont_limit_reg);
1950
1951 dct_offset_en = (u8) ((dct_cont_base_reg >> 3) & BIT(0));
1952 dct_sel = (u8) ((dct_cont_base_reg >> 4) & 0x7);
1953
1954 edac_dbg(1, "(range %d) SystemAddr= 0x%llx Limit=0x%llx\n",
1955 range, sys_addr, get_dram_limit(pvt, range));
1956
1957 if (!(get_dram_base(pvt, range) <= sys_addr) &&
1958 !(get_dram_limit(pvt, range) >= sys_addr))
1959 return -EINVAL;
1960
1961 if (dhar_valid(pvt) &&
1962 dhar_base(pvt) <= sys_addr &&
1963 sys_addr < BIT_64(32)) {
1964 amd64_warn("Huh? Address is in the MMIO hole: 0x%016llx\n",
1965 sys_addr);
1966 return -EINVAL;
1967 }
1968
1969 /* Verify sys_addr is within DCT Range. */
4fc06b31
AG
1970 dct_base = (u64) dct_sel_baseaddr(pvt);
1971 dct_limit = (dct_cont_limit_reg >> 11) & 0x1FFF;
18b94f66
AG
1972
1973 if (!(dct_cont_base_reg & BIT(0)) &&
4fc06b31
AG
1974 !(dct_base <= (sys_addr >> 27) &&
1975 dct_limit >= (sys_addr >> 27)))
18b94f66
AG
1976 return -EINVAL;
1977
1978 /* Verify number of dct's that participate in channel interleaving. */
1979 num_dcts_intlv = (int) hweight8(intlv_en);
1980
1981 if (!(num_dcts_intlv % 2 == 0) || (num_dcts_intlv > 4))
1982 return -EINVAL;
1983
dc0a50a8
YG
1984 if (pvt->model >= 0x60)
1985 channel = f1x_determine_channel(pvt, sys_addr, false, intlv_en);
1986 else
1987 channel = f15_m30h_determine_channel(pvt, sys_addr, intlv_en,
1988 num_dcts_intlv, dct_sel);
18b94f66
AG
1989
1990 /* Verify we stay within the MAX number of channels allowed */
7f3f5240 1991 if (channel > 3)
18b94f66
AG
1992 return -EINVAL;
1993
1994 leg_mmio_hole = (u8) (dct_cont_base_reg >> 1 & BIT(0));
1995
1996 /* Get normalized DCT addr */
1997 if (leg_mmio_hole && (sys_addr >= BIT_64(32)))
1998 chan_offset = dhar_offset;
1999 else
4fc06b31 2000 chan_offset = dct_base << 27;
18b94f66
AG
2001
2002 chan_addr = sys_addr - chan_offset;
2003
2004 /* remove channel interleave */
2005 if (num_dcts_intlv == 2) {
2006 if (intlv_addr == 0x4)
2007 chan_addr = ((chan_addr >> 9) << 8) |
2008 (chan_addr & 0xff);
2009 else if (intlv_addr == 0x5)
2010 chan_addr = ((chan_addr >> 10) << 9) |
2011 (chan_addr & 0x1ff);
2012 else
2013 return -EINVAL;
2014
2015 } else if (num_dcts_intlv == 4) {
2016 if (intlv_addr == 0x4)
2017 chan_addr = ((chan_addr >> 10) << 8) |
2018 (chan_addr & 0xff);
2019 else if (intlv_addr == 0x5)
2020 chan_addr = ((chan_addr >> 11) << 9) |
2021 (chan_addr & 0x1ff);
2022 else
2023 return -EINVAL;
2024 }
2025
2026 if (dct_offset_en) {
2027 amd64_read_pci_cfg(pvt->F1,
2028 DRAM_CONT_HIGH_OFF + (int) channel * 4,
2029 &tmp);
4fc06b31 2030 chan_addr += (u64) ((tmp >> 11) & 0xfff) << 27;
18b94f66
AG
2031 }
2032
2033 f15h_select_dct(pvt, channel);
2034
2035 edac_dbg(1, " Normalized DCT addr: 0x%llx\n", chan_addr);
2036
2037 /*
2038 * Find Chip select:
2039 * if channel = 3, then alias it to 1. This is because, in F15 M30h,
2040 * there is support for 4 DCT's, but only 2 are currently functional.
2041 * They are DCT0 and DCT3. But we have read all registers of DCT3 into
2042 * pvt->csels[1]. So we need to use '1' here to get correct info.
2043 * Refer F15 M30h BKDG Section 2.10 and 2.10.3 for clarifications.
2044 */
2045 alias_channel = (channel == 3) ? 1 : channel;
2046
2047 cs_found = f1x_lookup_addr_in_dct(chan_addr, node_id, alias_channel);
2048
2049 if (cs_found >= 0)
2050 *chan_sel = alias_channel;
2051
2052 return cs_found;
2053}
2054
2055static int f1x_translate_sysaddr_to_cs(struct amd64_pvt *pvt,
2056 u64 sys_addr,
2057 int *chan_sel)
f71d0a05 2058{
e761359a
BP
2059 int cs_found = -EINVAL;
2060 unsigned range;
f71d0a05 2061
7f19bf75 2062 for (range = 0; range < DRAM_RANGES; range++) {
7f19bf75 2063 if (!dram_rw(pvt, range))
f71d0a05
DT
2064 continue;
2065
18b94f66
AG
2066 if (pvt->fam == 0x15 && pvt->model >= 0x30)
2067 cs_found = f15_m30h_match_to_this_node(pvt, range,
2068 sys_addr,
2069 chan_sel);
f71d0a05 2070
18b94f66
AG
2071 else if ((get_dram_base(pvt, range) <= sys_addr) &&
2072 (get_dram_limit(pvt, range) >= sys_addr)) {
b15f0fca 2073 cs_found = f1x_match_to_this_node(pvt, range,
33ca0643 2074 sys_addr, chan_sel);
f71d0a05
DT
2075 if (cs_found >= 0)
2076 break;
2077 }
2078 }
2079 return cs_found;
2080}
2081
2082/*
bdc30a0c
BP
2083 * For reference see "2.8.5 Routing DRAM Requests" in F10 BKDG. This code maps
2084 * a @sys_addr to NodeID, DCT (channel) and chip select (CSROW).
f71d0a05 2085 *
bdc30a0c
BP
2086 * The @sys_addr is usually an error address received from the hardware
2087 * (MCX_ADDR).
f71d0a05 2088 */
b15f0fca 2089static void f1x_map_sysaddr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr,
33ca0643 2090 struct err_info *err)
f71d0a05
DT
2091{
2092 struct amd64_pvt *pvt = mci->pvt_info;
f71d0a05 2093
33ca0643 2094 error_address_to_page_and_offset(sys_addr, err);
ab5a503c 2095
33ca0643
BP
2096 err->csrow = f1x_translate_sysaddr_to_cs(pvt, sys_addr, &err->channel);
2097 if (err->csrow < 0) {
2098 err->err_code = ERR_CSROW;
bdc30a0c
BP
2099 return;
2100 }
2101
bdc30a0c
BP
2102 /*
2103 * We need the syndromes for channel detection only when we're
2104 * ganged. Otherwise @chan should already contain the channel at
2105 * this point.
2106 */
a97fa68e 2107 if (dct_ganging_enabled(pvt))
33ca0643 2108 err->channel = get_channel_from_ecc_syndrome(mci, err->syndrome);
f71d0a05
DT
2109}
2110
f71d0a05 2111/*
8566c4df 2112 * debug routine to display the memory sizes of all logical DIMMs and its
cb328507 2113 * CSROWs
f71d0a05 2114 */
d1ea71cd 2115static void debug_display_dimm_sizes(struct amd64_pvt *pvt, u8 ctrl)
f71d0a05 2116{
bb89f5a0 2117 int dimm, size0, size1;
525a1b20
BP
2118 u32 *dcsb = ctrl ? pvt->csels[1].csbases : pvt->csels[0].csbases;
2119 u32 dbam = ctrl ? pvt->dbam1 : pvt->dbam0;
f71d0a05 2120
a4b4bedc 2121 if (pvt->fam == 0xf) {
8566c4df 2122 /* K8 families < revF not supported yet */
1433eb99 2123 if (pvt->ext_model < K8_REV_F)
8566c4df
BP
2124 return;
2125 else
2126 WARN_ON(ctrl != 0);
2127 }
2128
7981a28f
AG
2129 if (pvt->fam == 0x10) {
2130 dbam = (ctrl && !dct_ganging_enabled(pvt)) ? pvt->dbam1
2131 : pvt->dbam0;
2132 dcsb = (ctrl && !dct_ganging_enabled(pvt)) ?
2133 pvt->csels[1].csbases :
2134 pvt->csels[0].csbases;
2135 } else if (ctrl) {
2136 dbam = pvt->dbam0;
2137 dcsb = pvt->csels[1].csbases;
2138 }
956b9ba1
JP
2139 edac_dbg(1, "F2x%d80 (DRAM Bank Address Mapping): 0x%08x\n",
2140 ctrl, dbam);
f71d0a05 2141
8566c4df
BP
2142 edac_printk(KERN_DEBUG, EDAC_MC, "DCT%d chip selects:\n", ctrl);
2143
f71d0a05
DT
2144 /* Dump memory sizes for DIMM and its CSROWs */
2145 for (dimm = 0; dimm < 4; dimm++) {
2146
2147 size0 = 0;
11c75ead 2148 if (dcsb[dimm*2] & DCSB_CS_ENABLE)
07ed82ef
YG
2149 /*
2150 * For F15m60h, we need multiplier for LRDIMM cs_size
2151 * calculation. We pass dimm value to the dbam_to_cs
a597d2a5
AG
2152 * mapper so we can find the multiplier from the
2153 * corresponding DCSM.
2154 */
41d8bfab 2155 size0 = pvt->ops->dbam_to_cs(pvt, ctrl,
a597d2a5
AG
2156 DBAM_DIMM(dimm, dbam),
2157 dimm);
f71d0a05
DT
2158
2159 size1 = 0;
11c75ead 2160 if (dcsb[dimm*2 + 1] & DCSB_CS_ENABLE)
41d8bfab 2161 size1 = pvt->ops->dbam_to_cs(pvt, ctrl,
a597d2a5
AG
2162 DBAM_DIMM(dimm, dbam),
2163 dimm);
f71d0a05 2164
24f9a7fe 2165 amd64_info(EDAC_MC ": %d: %5dMB %d: %5dMB\n",
bb89f5a0
BP
2166 dimm * 2, size0,
2167 dimm * 2 + 1, size1);
f71d0a05
DT
2168 }
2169}
2170
d1ea71cd 2171static struct amd64_family_type family_types[] = {
4d37607a 2172 [K8_CPUS] = {
0092b20d 2173 .ctl_name = "K8",
8d5b5d9c 2174 .f1_id = PCI_DEVICE_ID_AMD_K8_NB_ADDRMAP,
3f37a36b 2175 .f2_id = PCI_DEVICE_ID_AMD_K8_NB_MEMCTL,
4d37607a 2176 .ops = {
1433eb99 2177 .early_channel_count = k8_early_channel_count,
1433eb99
BP
2178 .map_sysaddr_to_csrow = k8_map_sysaddr_to_csrow,
2179 .dbam_to_cs = k8_dbam_to_chip_select,
4d37607a
DT
2180 }
2181 },
2182 [F10_CPUS] = {
0092b20d 2183 .ctl_name = "F10h",
8d5b5d9c 2184 .f1_id = PCI_DEVICE_ID_AMD_10H_NB_MAP,
3f37a36b 2185 .f2_id = PCI_DEVICE_ID_AMD_10H_NB_DRAM,
4d37607a 2186 .ops = {
7d20d14d 2187 .early_channel_count = f1x_early_channel_count,
b15f0fca 2188 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
1433eb99 2189 .dbam_to_cs = f10_dbam_to_chip_select,
b2b0c605
BP
2190 }
2191 },
2192 [F15_CPUS] = {
2193 .ctl_name = "F15h",
df71a053 2194 .f1_id = PCI_DEVICE_ID_AMD_15H_NB_F1,
3f37a36b 2195 .f2_id = PCI_DEVICE_ID_AMD_15H_NB_F2,
b2b0c605 2196 .ops = {
7d20d14d 2197 .early_channel_count = f1x_early_channel_count,
b15f0fca 2198 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
41d8bfab 2199 .dbam_to_cs = f15_dbam_to_chip_select,
4d37607a
DT
2200 }
2201 },
18b94f66
AG
2202 [F15_M30H_CPUS] = {
2203 .ctl_name = "F15h_M30h",
2204 .f1_id = PCI_DEVICE_ID_AMD_15H_M30H_NB_F1,
3f37a36b 2205 .f2_id = PCI_DEVICE_ID_AMD_15H_M30H_NB_F2,
18b94f66
AG
2206 .ops = {
2207 .early_channel_count = f1x_early_channel_count,
2208 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
2209 .dbam_to_cs = f16_dbam_to_chip_select,
18b94f66
AG
2210 }
2211 },
a597d2a5
AG
2212 [F15_M60H_CPUS] = {
2213 .ctl_name = "F15h_M60h",
2214 .f1_id = PCI_DEVICE_ID_AMD_15H_M60H_NB_F1,
3f37a36b 2215 .f2_id = PCI_DEVICE_ID_AMD_15H_M60H_NB_F2,
a597d2a5
AG
2216 .ops = {
2217 .early_channel_count = f1x_early_channel_count,
2218 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
2219 .dbam_to_cs = f15_m60h_dbam_to_chip_select,
2220 }
2221 },
94c1acf2
AG
2222 [F16_CPUS] = {
2223 .ctl_name = "F16h",
2224 .f1_id = PCI_DEVICE_ID_AMD_16H_NB_F1,
3f37a36b 2225 .f2_id = PCI_DEVICE_ID_AMD_16H_NB_F2,
94c1acf2
AG
2226 .ops = {
2227 .early_channel_count = f1x_early_channel_count,
2228 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
2229 .dbam_to_cs = f16_dbam_to_chip_select,
94c1acf2
AG
2230 }
2231 },
85a8885b
AG
2232 [F16_M30H_CPUS] = {
2233 .ctl_name = "F16h_M30h",
2234 .f1_id = PCI_DEVICE_ID_AMD_16H_M30H_NB_F1,
3f37a36b 2235 .f2_id = PCI_DEVICE_ID_AMD_16H_M30H_NB_F2,
85a8885b
AG
2236 .ops = {
2237 .early_channel_count = f1x_early_channel_count,
2238 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
2239 .dbam_to_cs = f16_dbam_to_chip_select,
85a8885b
AG
2240 }
2241 },
f1cbbec9
YG
2242 [F17_CPUS] = {
2243 .ctl_name = "F17h",
2244 .f0_id = PCI_DEVICE_ID_AMD_17H_DF_F0,
2245 .f6_id = PCI_DEVICE_ID_AMD_17H_DF_F6,
2246 .ops = {
2247 .early_channel_count = f17_early_channel_count,
2248 .dbam_to_cs = f17_base_addr_to_cs_size,
2249 }
2250 },
8960de4a
MJ
2251 [F17_M10H_CPUS] = {
2252 .ctl_name = "F17h_M10h",
2253 .f0_id = PCI_DEVICE_ID_AMD_17H_M10H_DF_F0,
2254 .f6_id = PCI_DEVICE_ID_AMD_17H_M10H_DF_F6,
2255 .ops = {
2256 .early_channel_count = f17_early_channel_count,
2257 .dbam_to_cs = f17_base_addr_to_cs_size,
2258 }
2259 },
6e846239
YG
2260 [F17_M30H_CPUS] = {
2261 .ctl_name = "F17h_M30h",
2262 .f0_id = PCI_DEVICE_ID_AMD_17H_M30H_DF_F0,
2263 .f6_id = PCI_DEVICE_ID_AMD_17H_M30H_DF_F6,
2264 .ops = {
2265 .early_channel_count = f17_early_channel_count,
2266 .dbam_to_cs = f17_base_addr_to_cs_size,
2267 }
2268 },
4d37607a
DT
2269};
2270
b1289d6f 2271/*
bfc04aec
BP
2272 * These are tables of eigenvectors (one per line) which can be used for the
2273 * construction of the syndrome tables. The modified syndrome search algorithm
2274 * uses those to find the symbol in error and thus the DIMM.
b1289d6f 2275 *
bfc04aec 2276 * Algorithm courtesy of Ross LaFetra from AMD.
b1289d6f 2277 */
c7e5301a 2278static const u16 x4_vectors[] = {
bfc04aec
BP
2279 0x2f57, 0x1afe, 0x66cc, 0xdd88,
2280 0x11eb, 0x3396, 0x7f4c, 0xeac8,
2281 0x0001, 0x0002, 0x0004, 0x0008,
2282 0x1013, 0x3032, 0x4044, 0x8088,
2283 0x106b, 0x30d6, 0x70fc, 0xe0a8,
2284 0x4857, 0xc4fe, 0x13cc, 0x3288,
2285 0x1ac5, 0x2f4a, 0x5394, 0xa1e8,
2286 0x1f39, 0x251e, 0xbd6c, 0x6bd8,
2287 0x15c1, 0x2a42, 0x89ac, 0x4758,
2288 0x2b03, 0x1602, 0x4f0c, 0xca08,
2289 0x1f07, 0x3a0e, 0x6b04, 0xbd08,
2290 0x8ba7, 0x465e, 0x244c, 0x1cc8,
2291 0x2b87, 0x164e, 0x642c, 0xdc18,
2292 0x40b9, 0x80de, 0x1094, 0x20e8,
2293 0x27db, 0x1eb6, 0x9dac, 0x7b58,
2294 0x11c1, 0x2242, 0x84ac, 0x4c58,
2295 0x1be5, 0x2d7a, 0x5e34, 0xa718,
2296 0x4b39, 0x8d1e, 0x14b4, 0x28d8,
2297 0x4c97, 0xc87e, 0x11fc, 0x33a8,
2298 0x8e97, 0x497e, 0x2ffc, 0x1aa8,
2299 0x16b3, 0x3d62, 0x4f34, 0x8518,
2300 0x1e2f, 0x391a, 0x5cac, 0xf858,
2301 0x1d9f, 0x3b7a, 0x572c, 0xfe18,
2302 0x15f5, 0x2a5a, 0x5264, 0xa3b8,
2303 0x1dbb, 0x3b66, 0x715c, 0xe3f8,
2304 0x4397, 0xc27e, 0x17fc, 0x3ea8,
2305 0x1617, 0x3d3e, 0x6464, 0xb8b8,
2306 0x23ff, 0x12aa, 0xab6c, 0x56d8,
2307 0x2dfb, 0x1ba6, 0x913c, 0x7328,
2308 0x185d, 0x2ca6, 0x7914, 0x9e28,
2309 0x171b, 0x3e36, 0x7d7c, 0xebe8,
2310 0x4199, 0x82ee, 0x19f4, 0x2e58,
2311 0x4807, 0xc40e, 0x130c, 0x3208,
2312 0x1905, 0x2e0a, 0x5804, 0xac08,
2313 0x213f, 0x132a, 0xadfc, 0x5ba8,
2314 0x19a9, 0x2efe, 0xb5cc, 0x6f88,
b1289d6f
DT
2315};
2316
c7e5301a 2317static const u16 x8_vectors[] = {
bfc04aec
BP
2318 0x0145, 0x028a, 0x2374, 0x43c8, 0xa1f0, 0x0520, 0x0a40, 0x1480,
2319 0x0211, 0x0422, 0x0844, 0x1088, 0x01b0, 0x44e0, 0x23c0, 0xed80,
2320 0x1011, 0x0116, 0x022c, 0x0458, 0x08b0, 0x8c60, 0x2740, 0x4e80,
2321 0x0411, 0x0822, 0x1044, 0x0158, 0x02b0, 0x2360, 0x46c0, 0xab80,
2322 0x0811, 0x1022, 0x012c, 0x0258, 0x04b0, 0x4660, 0x8cc0, 0x2780,
2323 0x2071, 0x40e2, 0xa0c4, 0x0108, 0x0210, 0x0420, 0x0840, 0x1080,
2324 0x4071, 0x80e2, 0x0104, 0x0208, 0x0410, 0x0820, 0x1040, 0x2080,
2325 0x8071, 0x0102, 0x0204, 0x0408, 0x0810, 0x1020, 0x2040, 0x4080,
2326 0x019d, 0x03d6, 0x136c, 0x2198, 0x50b0, 0xb2e0, 0x0740, 0x0e80,
2327 0x0189, 0x03ea, 0x072c, 0x0e58, 0x1cb0, 0x56e0, 0x37c0, 0xf580,
2328 0x01fd, 0x0376, 0x06ec, 0x0bb8, 0x1110, 0x2220, 0x4440, 0x8880,
2329 0x0163, 0x02c6, 0x1104, 0x0758, 0x0eb0, 0x2be0, 0x6140, 0xc280,
2330 0x02fd, 0x01c6, 0x0b5c, 0x1108, 0x07b0, 0x25a0, 0x8840, 0x6180,
2331 0x0801, 0x012e, 0x025c, 0x04b8, 0x1370, 0x26e0, 0x57c0, 0xb580,
2332 0x0401, 0x0802, 0x015c, 0x02b8, 0x22b0, 0x13e0, 0x7140, 0xe280,
2333 0x0201, 0x0402, 0x0804, 0x01b8, 0x11b0, 0x31a0, 0x8040, 0x7180,
2334 0x0101, 0x0202, 0x0404, 0x0808, 0x1010, 0x2020, 0x4040, 0x8080,
2335 0x0001, 0x0002, 0x0004, 0x0008, 0x0010, 0x0020, 0x0040, 0x0080,
2336 0x0100, 0x0200, 0x0400, 0x0800, 0x1000, 0x2000, 0x4000, 0x8000,
2337};
2338
c7e5301a 2339static int decode_syndrome(u16 syndrome, const u16 *vectors, unsigned num_vecs,
d34a6ecd 2340 unsigned v_dim)
b1289d6f 2341{
bfc04aec
BP
2342 unsigned int i, err_sym;
2343
2344 for (err_sym = 0; err_sym < num_vecs / v_dim; err_sym++) {
2345 u16 s = syndrome;
d34a6ecd
BP
2346 unsigned v_idx = err_sym * v_dim;
2347 unsigned v_end = (err_sym + 1) * v_dim;
bfc04aec
BP
2348
2349 /* walk over all 16 bits of the syndrome */
2350 for (i = 1; i < (1U << 16); i <<= 1) {
2351
2352 /* if bit is set in that eigenvector... */
2353 if (v_idx < v_end && vectors[v_idx] & i) {
2354 u16 ev_comp = vectors[v_idx++];
2355
2356 /* ... and bit set in the modified syndrome, */
2357 if (s & i) {
2358 /* remove it. */
2359 s ^= ev_comp;
4d37607a 2360
bfc04aec
BP
2361 if (!s)
2362 return err_sym;
2363 }
b1289d6f 2364
bfc04aec
BP
2365 } else if (s & i)
2366 /* can't get to zero, move to next symbol */
2367 break;
2368 }
b1289d6f
DT
2369 }
2370
956b9ba1 2371 edac_dbg(0, "syndrome(%x) not found\n", syndrome);
b1289d6f
DT
2372 return -1;
2373}
d27bf6fa 2374
bfc04aec
BP
2375static int map_err_sym_to_channel(int err_sym, int sym_size)
2376{
2377 if (sym_size == 4)
2378 switch (err_sym) {
2379 case 0x20:
2380 case 0x21:
2381 return 0;
2382 break;
2383 case 0x22:
2384 case 0x23:
2385 return 1;
2386 break;
2387 default:
2388 return err_sym >> 4;
2389 break;
2390 }
2391 /* x8 symbols */
2392 else
2393 switch (err_sym) {
2394 /* imaginary bits not in a DIMM */
2395 case 0x10:
2396 WARN(1, KERN_ERR "Invalid error symbol: 0x%x\n",
2397 err_sym);
2398 return -1;
2399 break;
2400
2401 case 0x11:
2402 return 0;
2403 break;
2404 case 0x12:
2405 return 1;
2406 break;
2407 default:
2408 return err_sym >> 3;
2409 break;
2410 }
2411 return -1;
2412}
2413
2414static int get_channel_from_ecc_syndrome(struct mem_ctl_info *mci, u16 syndrome)
2415{
2416 struct amd64_pvt *pvt = mci->pvt_info;
ad6a32e9
BP
2417 int err_sym = -1;
2418
a3b7db09 2419 if (pvt->ecc_sym_sz == 8)
ad6a32e9
BP
2420 err_sym = decode_syndrome(syndrome, x8_vectors,
2421 ARRAY_SIZE(x8_vectors),
a3b7db09
BP
2422 pvt->ecc_sym_sz);
2423 else if (pvt->ecc_sym_sz == 4)
ad6a32e9
BP
2424 err_sym = decode_syndrome(syndrome, x4_vectors,
2425 ARRAY_SIZE(x4_vectors),
a3b7db09 2426 pvt->ecc_sym_sz);
ad6a32e9 2427 else {
a3b7db09 2428 amd64_warn("Illegal syndrome type: %u\n", pvt->ecc_sym_sz);
ad6a32e9 2429 return err_sym;
bfc04aec 2430 }
ad6a32e9 2431
a3b7db09 2432 return map_err_sym_to_channel(err_sym, pvt->ecc_sym_sz);
bfc04aec
BP
2433}
2434
e70984d9 2435static void __log_ecc_error(struct mem_ctl_info *mci, struct err_info *err,
33ca0643 2436 u8 ecc_type)
d27bf6fa 2437{
33ca0643
BP
2438 enum hw_event_mc_err_type err_type;
2439 const char *string;
d27bf6fa 2440
33ca0643
BP
2441 if (ecc_type == 2)
2442 err_type = HW_EVENT_ERR_CORRECTED;
2443 else if (ecc_type == 1)
2444 err_type = HW_EVENT_ERR_UNCORRECTED;
d12a969e
YG
2445 else if (ecc_type == 3)
2446 err_type = HW_EVENT_ERR_DEFERRED;
33ca0643
BP
2447 else {
2448 WARN(1, "Something is rotten in the state of Denmark.\n");
d27bf6fa
DT
2449 return;
2450 }
2451
33ca0643
BP
2452 switch (err->err_code) {
2453 case DECODE_OK:
2454 string = "";
2455 break;
2456 case ERR_NODE:
2457 string = "Failed to map error addr to a node";
2458 break;
2459 case ERR_CSROW:
2460 string = "Failed to map error addr to a csrow";
2461 break;
2462 case ERR_CHANNEL:
713ad546
YG
2463 string = "Unknown syndrome - possible error reporting race";
2464 break;
2465 case ERR_SYND:
2466 string = "MCA_SYND not valid - unknown syndrome and csrow";
2467 break;
2468 case ERR_NORM_ADDR:
2469 string = "Cannot decode normalized address";
33ca0643
BP
2470 break;
2471 default:
2472 string = "WTF error";
2473 break;
d27bf6fa 2474 }
33ca0643
BP
2475
2476 edac_mc_handle_error(err_type, mci, 1,
2477 err->page, err->offset, err->syndrome,
2478 err->csrow, err->channel, -1,
2479 string, "");
d27bf6fa
DT
2480}
2481
df781d03 2482static inline void decode_bus_error(int node_id, struct mce *m)
d27bf6fa 2483{
0c510cc8
DB
2484 struct mem_ctl_info *mci;
2485 struct amd64_pvt *pvt;
f192c7b1 2486 u8 ecc_type = (m->status >> 45) & 0x3;
66fed2d4
BP
2487 u8 xec = XEC(m->status, 0x1f);
2488 u16 ec = EC(m->status);
33ca0643
BP
2489 u64 sys_addr;
2490 struct err_info err;
d27bf6fa 2491
0c510cc8
DB
2492 mci = edac_mc_find(node_id);
2493 if (!mci)
2494 return;
2495
2496 pvt = mci->pvt_info;
2497
66fed2d4 2498 /* Bail out early if this was an 'observed' error */
5980bb9c 2499 if (PP(ec) == NBSL_PP_OBS)
b70ef010 2500 return;
d27bf6fa 2501
ecaf5606
BP
2502 /* Do only ECC errors */
2503 if (xec && xec != F10_NBSL_EXT_ERR_ECC)
d27bf6fa 2504 return;
d27bf6fa 2505
33ca0643
BP
2506 memset(&err, 0, sizeof(err));
2507
a4b4bedc 2508 sys_addr = get_error_address(pvt, m);
33ca0643 2509
ecaf5606 2510 if (ecc_type == 2)
33ca0643
BP
2511 err.syndrome = extract_syndrome(m->status);
2512
2513 pvt->ops->map_sysaddr_to_csrow(mci, sys_addr, &err);
2514
e70984d9 2515 __log_ecc_error(mci, &err, ecc_type);
d27bf6fa
DT
2516}
2517
713ad546
YG
2518/*
2519 * To find the UMC channel represented by this bank we need to match on its
2520 * instance_id. The instance_id of a bank is held in the lower 32 bits of its
2521 * IPID.
bdcee774
YG
2522 *
2523 * Currently, we can derive the channel number by looking at the 6th nibble in
2524 * the instance_id. For example, instance_id=0xYXXXXX where Y is the channel
2525 * number.
713ad546 2526 */
bdcee774 2527static int find_umc_channel(struct mce *m)
713ad546 2528{
bdcee774 2529 return (m->ipid & GENMASK(31, 0)) >> 20;
713ad546
YG
2530}
2531
2532static void decode_umc_error(int node_id, struct mce *m)
2533{
2534 u8 ecc_type = (m->status >> 45) & 0x3;
2535 struct mem_ctl_info *mci;
2536 struct amd64_pvt *pvt;
2537 struct err_info err;
2538 u64 sys_addr;
2539
2540 mci = edac_mc_find(node_id);
2541 if (!mci)
2542 return;
2543
2544 pvt = mci->pvt_info;
2545
2546 memset(&err, 0, sizeof(err));
2547
2548 if (m->status & MCI_STATUS_DEFERRED)
2549 ecc_type = 3;
2550
bdcee774 2551 err.channel = find_umc_channel(m);
713ad546 2552
713ad546
YG
2553 if (!(m->status & MCI_STATUS_SYNDV)) {
2554 err.err_code = ERR_SYND;
2555 goto log_error;
2556 }
2557
2558 if (ecc_type == 2) {
2559 u8 length = (m->synd >> 18) & 0x3f;
2560
2561 if (length)
2562 err.syndrome = (m->synd >> 32) & GENMASK(length - 1, 0);
2563 else
2564 err.err_code = ERR_CHANNEL;
2565 }
2566
2567 err.csrow = m->synd & 0x7;
2568
3895ed6f
YG
2569 if (umc_normaddr_to_sysaddr(m->addr, pvt->mc_node_id, err.channel, &sys_addr)) {
2570 err.err_code = ERR_NORM_ADDR;
2571 goto log_error;
2572 }
2573
2574 error_address_to_page_and_offset(sys_addr, &err);
2575
713ad546
YG
2576log_error:
2577 __log_ecc_error(mci, &err, ecc_type);
2578}
2579
0ec449ee 2580/*
3f37a36b
BP
2581 * Use pvt->F3 which contains the F3 CPU PCI device to get the related
2582 * F1 (AddrMap) and F2 (Dct) devices. Return negative value on error.
936fc3af 2583 * Reserve F0 and F6 on systems with a UMC.
0ec449ee 2584 */
936fc3af
YG
2585static int
2586reserve_mc_sibling_devs(struct amd64_pvt *pvt, u16 pci_id1, u16 pci_id2)
2587{
2588 if (pvt->umc) {
2589 pvt->F0 = pci_get_related_function(pvt->F3->vendor, pci_id1, pvt->F3);
2590 if (!pvt->F0) {
5246c540 2591 amd64_err("F0 not found, device 0x%x (broken BIOS?)\n", pci_id1);
936fc3af
YG
2592 return -ENODEV;
2593 }
2594
2595 pvt->F6 = pci_get_related_function(pvt->F3->vendor, pci_id2, pvt->F3);
2596 if (!pvt->F6) {
2597 pci_dev_put(pvt->F0);
2598 pvt->F0 = NULL;
2599
5246c540 2600 amd64_err("F6 not found: device 0x%x (broken BIOS?)\n", pci_id2);
936fc3af
YG
2601 return -ENODEV;
2602 }
5246c540 2603
936fc3af
YG
2604 edac_dbg(1, "F0: %s\n", pci_name(pvt->F0));
2605 edac_dbg(1, "F3: %s\n", pci_name(pvt->F3));
2606 edac_dbg(1, "F6: %s\n", pci_name(pvt->F6));
2607
2608 return 0;
2609 }
2610
0ec449ee 2611 /* Reserve the ADDRESS MAP Device */
936fc3af 2612 pvt->F1 = pci_get_related_function(pvt->F3->vendor, pci_id1, pvt->F3);
8d5b5d9c 2613 if (!pvt->F1) {
5246c540 2614 amd64_err("F1 not found: device 0x%x (broken BIOS?)\n", pci_id1);
bbd0c1f6 2615 return -ENODEV;
0ec449ee
DT
2616 }
2617
3f37a36b 2618 /* Reserve the DCT Device */
936fc3af 2619 pvt->F2 = pci_get_related_function(pvt->F3->vendor, pci_id2, pvt->F3);
3f37a36b 2620 if (!pvt->F2) {
8d5b5d9c
BP
2621 pci_dev_put(pvt->F1);
2622 pvt->F1 = NULL;
0ec449ee 2623
5246c540
BP
2624 amd64_err("F2 not found: device 0x%x (broken BIOS?)\n", pci_id2);
2625 return -ENODEV;
0ec449ee 2626 }
936fc3af 2627
956b9ba1
JP
2628 edac_dbg(1, "F1: %s\n", pci_name(pvt->F1));
2629 edac_dbg(1, "F2: %s\n", pci_name(pvt->F2));
2630 edac_dbg(1, "F3: %s\n", pci_name(pvt->F3));
0ec449ee
DT
2631
2632 return 0;
2633}
2634
360b7f3c 2635static void free_mc_sibling_devs(struct amd64_pvt *pvt)
0ec449ee 2636{
936fc3af
YG
2637 if (pvt->umc) {
2638 pci_dev_put(pvt->F0);
2639 pci_dev_put(pvt->F6);
2640 } else {
2641 pci_dev_put(pvt->F1);
2642 pci_dev_put(pvt->F2);
2643 }
0ec449ee
DT
2644}
2645
b64ce7cd
YG
2646static void determine_ecc_sym_sz(struct amd64_pvt *pvt)
2647{
2648 pvt->ecc_sym_sz = 4;
2649
2650 if (pvt->umc) {
2651 u8 i;
2652
4d30d2bc 2653 for_each_umc(i) {
b64ce7cd 2654 /* Check enabled channels only: */
7835961d
YG
2655 if (pvt->umc[i].sdp_ctrl & UMC_SDP_INIT) {
2656 if (pvt->umc[i].ecc_ctrl & BIT(9)) {
2657 pvt->ecc_sym_sz = 16;
2658 return;
2659 } else if (pvt->umc[i].ecc_ctrl & BIT(7)) {
2660 pvt->ecc_sym_sz = 8;
2661 return;
2662 }
b64ce7cd
YG
2663 }
2664 }
7835961d 2665 } else if (pvt->fam >= 0x10) {
b64ce7cd
YG
2666 u32 tmp;
2667
2668 amd64_read_pci_cfg(pvt->F3, EXT_NB_MCA_CFG, &tmp);
2669 /* F16h has only DCT0, so no need to read dbam1. */
2670 if (pvt->fam != 0x16)
2671 amd64_read_dct_pci_cfg(pvt, 1, DBAM0, &pvt->dbam1);
2672
2673 /* F10h, revD and later can do x8 ECC too. */
2674 if ((pvt->fam > 0x10 || pvt->model > 7) && tmp & BIT(25))
2675 pvt->ecc_sym_sz = 8;
2676 }
2677}
2678
2679/*
2680 * Retrieve the hardware registers of the memory controller.
2681 */
2682static void __read_mc_regs_df(struct amd64_pvt *pvt)
2683{
2684 u8 nid = pvt->mc_node_id;
2685 struct amd64_umc *umc;
2686 u32 i, umc_base;
2687
2688 /* Read registers from each UMC */
4d30d2bc 2689 for_each_umc(i) {
b64ce7cd
YG
2690
2691 umc_base = get_umc_base(i);
2692 umc = &pvt->umc[i];
2693
07ed82ef
YG
2694 amd_smn_read(nid, umc_base + UMCCH_DIMM_CFG, &umc->dimm_cfg);
2695 amd_smn_read(nid, umc_base + UMCCH_UMC_CFG, &umc->umc_cfg);
b64ce7cd
YG
2696 amd_smn_read(nid, umc_base + UMCCH_SDP_CTRL, &umc->sdp_ctrl);
2697 amd_smn_read(nid, umc_base + UMCCH_ECC_CTRL, &umc->ecc_ctrl);
07ed82ef 2698 amd_smn_read(nid, umc_base + UMCCH_UMC_CAP_HI, &umc->umc_cap_hi);
b64ce7cd
YG
2699 }
2700}
2701
0ec449ee
DT
2702/*
2703 * Retrieve the hardware registers of the memory controller (this includes the
2704 * 'Address Map' and 'Misc' device regs)
2705 */
360b7f3c 2706static void read_mc_regs(struct amd64_pvt *pvt)
0ec449ee 2707{
b64ce7cd 2708 unsigned int range;
0ec449ee 2709 u64 msr_val;
0ec449ee
DT
2710
2711 /*
2712 * Retrieve TOP_MEM and TOP_MEM2; no masking off of reserved bits since
b64ce7cd 2713 * those are Read-As-Zero.
0ec449ee 2714 */
e97f8bb8 2715 rdmsrl(MSR_K8_TOP_MEM1, pvt->top_mem);
956b9ba1 2716 edac_dbg(0, " TOP_MEM: 0x%016llx\n", pvt->top_mem);
0ec449ee 2717
b64ce7cd 2718 /* Check first whether TOP_MEM2 is enabled: */
0ec449ee 2719 rdmsrl(MSR_K8_SYSCFG, msr_val);
b64ce7cd 2720 if (msr_val & BIT(21)) {
e97f8bb8 2721 rdmsrl(MSR_K8_TOP_MEM2, pvt->top_mem2);
956b9ba1 2722 edac_dbg(0, " TOP_MEM2: 0x%016llx\n", pvt->top_mem2);
b64ce7cd 2723 } else {
956b9ba1 2724 edac_dbg(0, " TOP_MEM2 disabled\n");
b64ce7cd
YG
2725 }
2726
2727 if (pvt->umc) {
2728 __read_mc_regs_df(pvt);
2729 amd64_read_pci_cfg(pvt->F0, DF_DHAR, &pvt->dhar);
2730
2731 goto skip;
2732 }
0ec449ee 2733
5980bb9c 2734 amd64_read_pci_cfg(pvt->F3, NBCAP, &pvt->nbcap);
0ec449ee 2735
5a5d2371 2736 read_dram_ctl_register(pvt);
0ec449ee 2737
7f19bf75
BP
2738 for (range = 0; range < DRAM_RANGES; range++) {
2739 u8 rw;
0ec449ee 2740
7f19bf75
BP
2741 /* read settings for this DRAM range */
2742 read_dram_base_limit_regs(pvt, range);
2743
2744 rw = dram_rw(pvt, range);
2745 if (!rw)
2746 continue;
2747
956b9ba1
JP
2748 edac_dbg(1, " DRAM range[%d], base: 0x%016llx; limit: 0x%016llx\n",
2749 range,
2750 get_dram_base(pvt, range),
2751 get_dram_limit(pvt, range));
7f19bf75 2752
956b9ba1
JP
2753 edac_dbg(1, " IntlvEn=%s; Range access: %s%s IntlvSel=%d DstNode=%d\n",
2754 dram_intlv_en(pvt, range) ? "Enabled" : "Disabled",
2755 (rw & 0x1) ? "R" : "-",
2756 (rw & 0x2) ? "W" : "-",
2757 dram_intlv_sel(pvt, range),
2758 dram_dst_node(pvt, range));
0ec449ee
DT
2759 }
2760
bc21fa57 2761 amd64_read_pci_cfg(pvt->F1, DHAR, &pvt->dhar);
7981a28f 2762 amd64_read_dct_pci_cfg(pvt, 0, DBAM0, &pvt->dbam0);
0ec449ee 2763
8d5b5d9c 2764 amd64_read_pci_cfg(pvt->F3, F10_ONLINE_SPARE, &pvt->online_spare);
0ec449ee 2765
7981a28f
AG
2766 amd64_read_dct_pci_cfg(pvt, 0, DCLR0, &pvt->dclr0);
2767 amd64_read_dct_pci_cfg(pvt, 0, DCHR0, &pvt->dchr0);
0ec449ee 2768
78da121e 2769 if (!dct_ganging_enabled(pvt)) {
7981a28f
AG
2770 amd64_read_dct_pci_cfg(pvt, 1, DCLR0, &pvt->dclr1);
2771 amd64_read_dct_pci_cfg(pvt, 1, DCHR0, &pvt->dchr1);
0ec449ee 2772 }
ad6a32e9 2773
b64ce7cd
YG
2774skip:
2775 read_dct_base_mask(pvt);
2776
a597d2a5
AG
2777 determine_memory_type(pvt);
2778 edac_dbg(1, " DIMM type: %s\n", edac_mem_types[pvt->dram_type]);
a3b7db09 2779
b64ce7cd 2780 determine_ecc_sym_sz(pvt);
ad6a32e9 2781
b2b0c605 2782 dump_misc_regs(pvt);
0ec449ee
DT
2783}
2784
2785/*
2786 * NOTE: CPU Revision Dependent code
2787 *
2788 * Input:
11c75ead 2789 * @csrow_nr ChipSelect Row Number (0..NUM_CHIPSELECTS-1)
0ec449ee
DT
2790 * k8 private pointer to -->
2791 * DRAM Bank Address mapping register
2792 * node_id
2793 * DCL register where dual_channel_active is
2794 *
2795 * The DBAM register consists of 4 sets of 4 bits each definitions:
2796 *
2797 * Bits: CSROWs
2798 * 0-3 CSROWs 0 and 1
2799 * 4-7 CSROWs 2 and 3
2800 * 8-11 CSROWs 4 and 5
2801 * 12-15 CSROWs 6 and 7
2802 *
2803 * Values range from: 0 to 15
2804 * The meaning of the values depends on CPU revision and dual-channel state,
2805 * see relevant BKDG more info.
2806 *
2807 * The memory controller provides for total of only 8 CSROWs in its current
2808 * architecture. Each "pair" of CSROWs normally represents just one DIMM in
2809 * single channel or two (2) DIMMs in dual channel mode.
2810 *
2811 * The following code logic collapses the various tables for CSROW based on CPU
2812 * revision.
2813 *
2814 * Returns:
2815 * The number of PAGE_SIZE pages on the specified CSROW number it
2816 * encompasses
2817 *
2818 */
eb77e6b8 2819static u32 get_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr_orig)
0ec449ee 2820{
f92cae45 2821 u32 dbam = dct ? pvt->dbam1 : pvt->dbam0;
eb77e6b8
YG
2822 int csrow_nr = csrow_nr_orig;
2823 u32 cs_mode, nr_pages;
0ec449ee 2824
eb77e6b8
YG
2825 if (!pvt->umc)
2826 csrow_nr >>= 1;
10de6497 2827
eb77e6b8 2828 cs_mode = DBAM_DIMM(csrow_nr, dbam);
0ec449ee 2829
eb77e6b8
YG
2830 nr_pages = pvt->ops->dbam_to_cs(pvt, dct, cs_mode, csrow_nr);
2831 nr_pages <<= 20 - PAGE_SHIFT;
0ec449ee 2832
10de6497 2833 edac_dbg(0, "csrow: %d, channel: %d, DBAM idx: %d\n",
eb77e6b8 2834 csrow_nr_orig, dct, cs_mode);
10de6497 2835 edac_dbg(0, "nr_pages/channel: %u\n", nr_pages);
0ec449ee
DT
2836
2837 return nr_pages;
2838}
2839
2840/*
2841 * Initialize the array of csrow attribute instances, based on the values
2842 * from pci config hardware registers.
2843 */
360b7f3c 2844static int init_csrows(struct mem_ctl_info *mci)
0ec449ee 2845{
10de6497 2846 struct amd64_pvt *pvt = mci->pvt_info;
2d09d8f3 2847 enum edac_type edac_mode = EDAC_NONE;
0ec449ee 2848 struct csrow_info *csrow;
de3910eb 2849 struct dimm_info *dimm;
10de6497 2850 int i, j, empty = 1;
a895bf8b 2851 int nr_pages = 0;
10de6497 2852 u32 val;
0ec449ee 2853
2d09d8f3
YG
2854 if (!pvt->umc) {
2855 amd64_read_pci_cfg(pvt->F3, NBCFG, &val);
0ec449ee 2856
2d09d8f3 2857 pvt->nbcfg = val;
0ec449ee 2858
2d09d8f3
YG
2859 edac_dbg(0, "node %d, NBCFG=0x%08x[ChipKillEccCap: %d|DramEccEn: %d]\n",
2860 pvt->mc_node_id, val,
2861 !!(val & NBCFG_CHIPKILL), !!(val & NBCFG_ECC_ENABLE));
2862 }
0ec449ee 2863
10de6497
BP
2864 /*
2865 * We iterate over DCT0 here but we look at DCT1 in parallel, if needed.
2866 */
11c75ead 2867 for_each_chip_select(i, 0, pvt) {
10de6497
BP
2868 bool row_dct0 = !!csrow_enabled(i, 0, pvt);
2869 bool row_dct1 = false;
0ec449ee 2870
a4b4bedc 2871 if (pvt->fam != 0xf)
10de6497
BP
2872 row_dct1 = !!csrow_enabled(i, 1, pvt);
2873
2874 if (!row_dct0 && !row_dct1)
0ec449ee 2875 continue;
0ec449ee 2876
10de6497 2877 csrow = mci->csrows[i];
0ec449ee 2878 empty = 0;
10de6497
BP
2879
2880 edac_dbg(1, "MC node: %d, csrow: %d\n",
2881 pvt->mc_node_id, i);
2882
1eef1282 2883 if (row_dct0) {
d1ea71cd 2884 nr_pages = get_csrow_nr_pages(pvt, 0, i);
1eef1282
MCC
2885 csrow->channels[0]->dimm->nr_pages = nr_pages;
2886 }
11c75ead 2887
10de6497 2888 /* K8 has only one DCT */
a4b4bedc 2889 if (pvt->fam != 0xf && row_dct1) {
d1ea71cd 2890 int row_dct1_pages = get_csrow_nr_pages(pvt, 1, i);
1eef1282
MCC
2891
2892 csrow->channels[1]->dimm->nr_pages = row_dct1_pages;
2893 nr_pages += row_dct1_pages;
2894 }
0ec449ee 2895
10de6497 2896 edac_dbg(1, "Total csrow%d pages: %u\n", i, nr_pages);
0ec449ee 2897
2d09d8f3
YG
2898 /* Determine DIMM ECC mode: */
2899 if (pvt->umc) {
2900 if (mci->edac_ctl_cap & EDAC_FLAG_S4ECD4ED)
2901 edac_mode = EDAC_S4ECD4ED;
2902 else if (mci->edac_ctl_cap & EDAC_FLAG_SECDED)
2903 edac_mode = EDAC_SECDED;
2904
2905 } else if (pvt->nbcfg & NBCFG_ECC_ENABLE) {
2906 edac_mode = (pvt->nbcfg & NBCFG_CHIPKILL)
2907 ? EDAC_S4ECD4ED
2908 : EDAC_SECDED;
2909 }
084a4fcc
MCC
2910
2911 for (j = 0; j < pvt->channel_count; j++) {
de3910eb 2912 dimm = csrow->channels[j]->dimm;
a597d2a5 2913 dimm->mtype = pvt->dram_type;
de3910eb 2914 dimm->edac_mode = edac_mode;
084a4fcc 2915 }
0ec449ee
DT
2916 }
2917
2918 return empty;
2919}
d27bf6fa 2920
f6d6ae96 2921/* get all cores on this DCT */
8b84c8df 2922static void get_cpus_on_this_dct_cpumask(struct cpumask *mask, u16 nid)
f6d6ae96
BP
2923{
2924 int cpu;
2925
2926 for_each_online_cpu(cpu)
2927 if (amd_get_nb_id(cpu) == nid)
2928 cpumask_set_cpu(cpu, mask);
2929}
2930
2931/* check MCG_CTL on all the cpus on this node */
d1ea71cd 2932static bool nb_mce_bank_enabled_on_node(u16 nid)
f6d6ae96
BP
2933{
2934 cpumask_var_t mask;
50542251 2935 int cpu, nbe;
f6d6ae96
BP
2936 bool ret = false;
2937
2938 if (!zalloc_cpumask_var(&mask, GFP_KERNEL)) {
24f9a7fe 2939 amd64_warn("%s: Error allocating mask\n", __func__);
f6d6ae96
BP
2940 return false;
2941 }
2942
2943 get_cpus_on_this_dct_cpumask(mask, nid);
2944
f6d6ae96
BP
2945 rdmsr_on_cpus(mask, MSR_IA32_MCG_CTL, msrs);
2946
2947 for_each_cpu(cpu, mask) {
50542251 2948 struct msr *reg = per_cpu_ptr(msrs, cpu);
5980bb9c 2949 nbe = reg->l & MSR_MCGCTL_NBE;
f6d6ae96 2950
956b9ba1
JP
2951 edac_dbg(0, "core: %u, MCG_CTL: 0x%llx, NB MSR is %s\n",
2952 cpu, reg->q,
2953 (nbe ? "enabled" : "disabled"));
f6d6ae96
BP
2954
2955 if (!nbe)
2956 goto out;
f6d6ae96
BP
2957 }
2958 ret = true;
2959
2960out:
f6d6ae96
BP
2961 free_cpumask_var(mask);
2962 return ret;
2963}
2964
c7e5301a 2965static int toggle_ecc_err_reporting(struct ecc_settings *s, u16 nid, bool on)
f6d6ae96
BP
2966{
2967 cpumask_var_t cmask;
50542251 2968 int cpu;
f6d6ae96
BP
2969
2970 if (!zalloc_cpumask_var(&cmask, GFP_KERNEL)) {
24f9a7fe 2971 amd64_warn("%s: error allocating mask\n", __func__);
0de27884 2972 return -ENOMEM;
f6d6ae96
BP
2973 }
2974
ae7bb7c6 2975 get_cpus_on_this_dct_cpumask(cmask, nid);
f6d6ae96 2976
f6d6ae96
BP
2977 rdmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs);
2978
2979 for_each_cpu(cpu, cmask) {
2980
50542251
BP
2981 struct msr *reg = per_cpu_ptr(msrs, cpu);
2982
f6d6ae96 2983 if (on) {
5980bb9c 2984 if (reg->l & MSR_MCGCTL_NBE)
ae7bb7c6 2985 s->flags.nb_mce_enable = 1;
f6d6ae96 2986
5980bb9c 2987 reg->l |= MSR_MCGCTL_NBE;
f6d6ae96
BP
2988 } else {
2989 /*
d95cf4de 2990 * Turn off NB MCE reporting only when it was off before
f6d6ae96 2991 */
ae7bb7c6 2992 if (!s->flags.nb_mce_enable)
5980bb9c 2993 reg->l &= ~MSR_MCGCTL_NBE;
f6d6ae96 2994 }
f6d6ae96
BP
2995 }
2996 wrmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs);
2997
f6d6ae96
BP
2998 free_cpumask_var(cmask);
2999
3000 return 0;
3001}
3002
c7e5301a 3003static bool enable_ecc_error_reporting(struct ecc_settings *s, u16 nid,
2299ef71 3004 struct pci_dev *F3)
f9431992 3005{
2299ef71 3006 bool ret = true;
c9f4f26e 3007 u32 value, mask = 0x3; /* UECC/CECC enable */
f9431992 3008
2299ef71
BP
3009 if (toggle_ecc_err_reporting(s, nid, ON)) {
3010 amd64_warn("Error enabling ECC reporting over MCGCTL!\n");
3011 return false;
3012 }
3013
c9f4f26e 3014 amd64_read_pci_cfg(F3, NBCTL, &value);
f9431992 3015
ae7bb7c6
BP
3016 s->old_nbctl = value & mask;
3017 s->nbctl_valid = true;
f9431992
DT
3018
3019 value |= mask;
c9f4f26e 3020 amd64_write_pci_cfg(F3, NBCTL, value);
f9431992 3021
a97fa68e 3022 amd64_read_pci_cfg(F3, NBCFG, &value);
f9431992 3023
956b9ba1
JP
3024 edac_dbg(0, "1: node %d, NBCFG=0x%08x[DramEccEn: %d]\n",
3025 nid, value, !!(value & NBCFG_ECC_ENABLE));
f9431992 3026
a97fa68e 3027 if (!(value & NBCFG_ECC_ENABLE)) {
24f9a7fe 3028 amd64_warn("DRAM ECC disabled on this node, enabling...\n");
f9431992 3029
ae7bb7c6 3030 s->flags.nb_ecc_prev = 0;
d95cf4de 3031
f9431992 3032 /* Attempt to turn on DRAM ECC Enable */
a97fa68e
BP
3033 value |= NBCFG_ECC_ENABLE;
3034 amd64_write_pci_cfg(F3, NBCFG, value);
f9431992 3035
a97fa68e 3036 amd64_read_pci_cfg(F3, NBCFG, &value);
f9431992 3037
a97fa68e 3038 if (!(value & NBCFG_ECC_ENABLE)) {
24f9a7fe
BP
3039 amd64_warn("Hardware rejected DRAM ECC enable,"
3040 "check memory DIMM configuration.\n");
2299ef71 3041 ret = false;
f9431992 3042 } else {
24f9a7fe 3043 amd64_info("Hardware accepted DRAM ECC Enable\n");
f9431992 3044 }
d95cf4de 3045 } else {
ae7bb7c6 3046 s->flags.nb_ecc_prev = 1;
f9431992 3047 }
d95cf4de 3048
956b9ba1
JP
3049 edac_dbg(0, "2: node %d, NBCFG=0x%08x[DramEccEn: %d]\n",
3050 nid, value, !!(value & NBCFG_ECC_ENABLE));
f9431992 3051
2299ef71 3052 return ret;
f9431992
DT
3053}
3054
c7e5301a 3055static void restore_ecc_error_reporting(struct ecc_settings *s, u16 nid,
360b7f3c 3056 struct pci_dev *F3)
f9431992 3057{
c9f4f26e
BP
3058 u32 value, mask = 0x3; /* UECC/CECC enable */
3059
ae7bb7c6 3060 if (!s->nbctl_valid)
f9431992
DT
3061 return;
3062
c9f4f26e 3063 amd64_read_pci_cfg(F3, NBCTL, &value);
f9431992 3064 value &= ~mask;
ae7bb7c6 3065 value |= s->old_nbctl;
f9431992 3066
c9f4f26e 3067 amd64_write_pci_cfg(F3, NBCTL, value);
f9431992 3068
ae7bb7c6
BP
3069 /* restore previous BIOS DRAM ECC "off" setting we force-enabled */
3070 if (!s->flags.nb_ecc_prev) {
a97fa68e
BP
3071 amd64_read_pci_cfg(F3, NBCFG, &value);
3072 value &= ~NBCFG_ECC_ENABLE;
3073 amd64_write_pci_cfg(F3, NBCFG, value);
d95cf4de
BP
3074 }
3075
3076 /* restore the NB Enable MCGCTL bit */
2299ef71 3077 if (toggle_ecc_err_reporting(s, nid, OFF))
24f9a7fe 3078 amd64_warn("Error restoring NB MCGCTL settings!\n");
f9431992
DT
3079}
3080
3081/*
2299ef71
BP
3082 * EDAC requires that the BIOS have ECC enabled before
3083 * taking over the processing of ECC errors. A command line
3084 * option allows to force-enable hardware ECC later in
3085 * enable_ecc_error_reporting().
f9431992 3086 */
cab4d277
BP
3087static const char *ecc_msg =
3088 "ECC disabled in the BIOS or no ECC capability, module will not load.\n"
3089 " Either enable ECC checking or force module loading by setting "
3090 "'ecc_enable_override'.\n"
3091 " (Note that use of the override may cause unknown side effects.)\n";
be3468e8 3092
c7e5301a 3093static bool ecc_enabled(struct pci_dev *F3, u16 nid)
f9431992 3094{
06724535 3095 bool nb_mce_en = false;
196b79fc
YG
3096 u8 ecc_en = 0, i;
3097 u32 value;
f9431992 3098
196b79fc
YG
3099 if (boot_cpu_data.x86 >= 0x17) {
3100 u8 umc_en_mask = 0, ecc_en_mask = 0;
f9431992 3101
4d30d2bc 3102 for_each_umc(i) {
196b79fc
YG
3103 u32 base = get_umc_base(i);
3104
3105 /* Only check enabled UMCs. */
3106 if (amd_smn_read(nid, base + UMCCH_SDP_CTRL, &value))
3107 continue;
3108
3109 if (!(value & UMC_SDP_INIT))
3110 continue;
3111
3112 umc_en_mask |= BIT(i);
3113
3114 if (amd_smn_read(nid, base + UMCCH_UMC_CAP_HI, &value))
3115 continue;
3116
3117 if (value & UMC_ECC_ENABLED)
3118 ecc_en_mask |= BIT(i);
3119 }
3120
3121 /* Check whether at least one UMC is enabled: */
3122 if (umc_en_mask)
3123 ecc_en = umc_en_mask == ecc_en_mask;
11ab1cae
YG
3124 else
3125 edac_dbg(0, "Node %d: No enabled UMCs.\n", nid);
196b79fc
YG
3126
3127 /* Assume UMC MCA banks are enabled. */
3128 nb_mce_en = true;
3129 } else {
3130 amd64_read_pci_cfg(F3, NBCFG, &value);
f9431992 3131
196b79fc
YG
3132 ecc_en = !!(value & NBCFG_ECC_ENABLE);
3133
3134 nb_mce_en = nb_mce_bank_enabled_on_node(nid);
3135 if (!nb_mce_en)
11ab1cae 3136 edac_dbg(0, "NB MCE bank disabled, set MSR 0x%08x[4] on node %d to enable.\n",
196b79fc
YG
3137 MSR_IA32_MCG_CTL, nid);
3138 }
3139
11ab1cae
YG
3140 amd64_info("Node %d: DRAM ECC %s.\n",
3141 nid, (ecc_en ? "enabled" : "disabled"));
f9431992 3142
2299ef71 3143 if (!ecc_en || !nb_mce_en) {
11ab1cae 3144 amd64_info("%s", ecc_msg);
2299ef71
BP
3145 return false;
3146 }
3147 return true;
f9431992
DT
3148}
3149
2d09d8f3
YG
3150static inline void
3151f17h_determine_edac_ctl_cap(struct mem_ctl_info *mci, struct amd64_pvt *pvt)
3152{
662877e7 3153 u8 i, ecc_en = 1, cpk_en = 1, dev_x4 = 1, dev_x16 = 1;
2d09d8f3 3154
4d30d2bc 3155 for_each_umc(i) {
2d09d8f3
YG
3156 if (pvt->umc[i].sdp_ctrl & UMC_SDP_INIT) {
3157 ecc_en &= !!(pvt->umc[i].umc_cap_hi & UMC_ECC_ENABLED);
3158 cpk_en &= !!(pvt->umc[i].umc_cap_hi & UMC_ECC_CHIPKILL_CAP);
662877e7
YG
3159
3160 dev_x4 &= !!(pvt->umc[i].dimm_cfg & BIT(6));
3161 dev_x16 &= !!(pvt->umc[i].dimm_cfg & BIT(7));
2d09d8f3
YG
3162 }
3163 }
3164
3165 /* Set chipkill only if ECC is enabled: */
3166 if (ecc_en) {
3167 mci->edac_ctl_cap |= EDAC_FLAG_SECDED;
3168
662877e7
YG
3169 if (!cpk_en)
3170 return;
3171
3172 if (dev_x4)
2d09d8f3 3173 mci->edac_ctl_cap |= EDAC_FLAG_S4ECD4ED;
662877e7
YG
3174 else if (dev_x16)
3175 mci->edac_ctl_cap |= EDAC_FLAG_S16ECD16ED;
3176 else
3177 mci->edac_ctl_cap |= EDAC_FLAG_S8ECD8ED;
2d09d8f3
YG
3178 }
3179}
3180
df71a053
BP
3181static void setup_mci_misc_attrs(struct mem_ctl_info *mci,
3182 struct amd64_family_type *fam)
7d6034d3
DT
3183{
3184 struct amd64_pvt *pvt = mci->pvt_info;
3185
3186 mci->mtype_cap = MEM_FLAG_DDR2 | MEM_FLAG_RDDR2;
3187 mci->edac_ctl_cap = EDAC_FLAG_NONE;
7d6034d3 3188
2d09d8f3
YG
3189 if (pvt->umc) {
3190 f17h_determine_edac_ctl_cap(mci, pvt);
3191 } else {
3192 if (pvt->nbcap & NBCAP_SECDED)
3193 mci->edac_ctl_cap |= EDAC_FLAG_SECDED;
7d6034d3 3194
2d09d8f3
YG
3195 if (pvt->nbcap & NBCAP_CHIPKILL)
3196 mci->edac_ctl_cap |= EDAC_FLAG_S4ECD4ED;
3197 }
7d6034d3 3198
d1ea71cd 3199 mci->edac_cap = determine_edac_cap(pvt);
7d6034d3 3200 mci->mod_name = EDAC_MOD_STR;
df71a053 3201 mci->ctl_name = fam->ctl_name;
e7934b70 3202 mci->dev_name = pci_name(pvt->F3);
7d6034d3
DT
3203 mci->ctl_page_to_phys = NULL;
3204
7d6034d3 3205 /* memory scrubber interface */
d1ea71cd
BP
3206 mci->set_sdram_scrub_rate = set_scrub_rate;
3207 mci->get_sdram_scrub_rate = get_scrub_rate;
7d6034d3
DT
3208}
3209
0092b20d
BP
3210/*
3211 * returns a pointer to the family descriptor on success, NULL otherwise.
3212 */
d1ea71cd 3213static struct amd64_family_type *per_family_init(struct amd64_pvt *pvt)
395ae783 3214{
0092b20d
BP
3215 struct amd64_family_type *fam_type = NULL;
3216
18b94f66 3217 pvt->ext_model = boot_cpu_data.x86_model >> 4;
b399151c 3218 pvt->stepping = boot_cpu_data.x86_stepping;
18b94f66
AG
3219 pvt->model = boot_cpu_data.x86_model;
3220 pvt->fam = boot_cpu_data.x86;
3221
3222 switch (pvt->fam) {
395ae783 3223 case 0xf:
d1ea71cd
BP
3224 fam_type = &family_types[K8_CPUS];
3225 pvt->ops = &family_types[K8_CPUS].ops;
395ae783 3226 break;
df71a053 3227
395ae783 3228 case 0x10:
d1ea71cd
BP
3229 fam_type = &family_types[F10_CPUS];
3230 pvt->ops = &family_types[F10_CPUS].ops;
df71a053
BP
3231 break;
3232
3233 case 0x15:
18b94f66 3234 if (pvt->model == 0x30) {
d1ea71cd
BP
3235 fam_type = &family_types[F15_M30H_CPUS];
3236 pvt->ops = &family_types[F15_M30H_CPUS].ops;
18b94f66 3237 break;
a597d2a5
AG
3238 } else if (pvt->model == 0x60) {
3239 fam_type = &family_types[F15_M60H_CPUS];
3240 pvt->ops = &family_types[F15_M60H_CPUS].ops;
3241 break;
18b94f66
AG
3242 }
3243
d1ea71cd
BP
3244 fam_type = &family_types[F15_CPUS];
3245 pvt->ops = &family_types[F15_CPUS].ops;
395ae783
BP
3246 break;
3247
94c1acf2 3248 case 0x16:
85a8885b
AG
3249 if (pvt->model == 0x30) {
3250 fam_type = &family_types[F16_M30H_CPUS];
3251 pvt->ops = &family_types[F16_M30H_CPUS].ops;
3252 break;
3253 }
d1ea71cd
BP
3254 fam_type = &family_types[F16_CPUS];
3255 pvt->ops = &family_types[F16_CPUS].ops;
94c1acf2
AG
3256 break;
3257
f1cbbec9 3258 case 0x17:
8960de4a
MJ
3259 if (pvt->model >= 0x10 && pvt->model <= 0x2f) {
3260 fam_type = &family_types[F17_M10H_CPUS];
3261 pvt->ops = &family_types[F17_M10H_CPUS].ops;
3262 break;
6e846239
YG
3263 } else if (pvt->model >= 0x30 && pvt->model <= 0x3f) {
3264 fam_type = &family_types[F17_M30H_CPUS];
3265 pvt->ops = &family_types[F17_M30H_CPUS].ops;
3266 break;
8960de4a 3267 }
c4a3e946
PW
3268 /* fall through */
3269 case 0x18:
f1cbbec9
YG
3270 fam_type = &family_types[F17_CPUS];
3271 pvt->ops = &family_types[F17_CPUS].ops;
c4a3e946
PW
3272
3273 if (pvt->fam == 0x18)
3274 family_types[F17_CPUS].ctl_name = "F18h";
f1cbbec9
YG
3275 break;
3276
395ae783 3277 default:
24f9a7fe 3278 amd64_err("Unsupported family!\n");
0092b20d 3279 return NULL;
395ae783 3280 }
0092b20d 3281
df71a053 3282 amd64_info("%s %sdetected (node %d).\n", fam_type->ctl_name,
18b94f66 3283 (pvt->fam == 0xf ?
24f9a7fe
BP
3284 (pvt->ext_model >= K8_REV_F ? "revF or later "
3285 : "revE or earlier ")
3286 : ""), pvt->mc_node_id);
0092b20d 3287 return fam_type;
395ae783
BP
3288}
3289
e339f1ec
TI
3290static const struct attribute_group *amd64_edac_attr_groups[] = {
3291#ifdef CONFIG_EDAC_DEBUG
3292 &amd64_edac_dbg_group,
3293#endif
3294#ifdef CONFIG_EDAC_AMD64_ERROR_INJECTION
3295 &amd64_edac_inj_group,
3296#endif
3297 NULL
3298};
3299
bdcee774
YG
3300/* Set the number of Unified Memory Controllers in the system. */
3301static void compute_num_umcs(void)
3302{
3303 u8 model = boot_cpu_data.x86_model;
3304
3305 if (boot_cpu_data.x86 < 0x17)
3306 return;
3307
3308 if (model >= 0x30 && model <= 0x3f)
3309 num_umcs = 8;
3310 else
3311 num_umcs = 2;
3312
3313 edac_dbg(1, "Number of UMCs: %x", num_umcs);
3314}
3315
3f37a36b 3316static int init_one_instance(unsigned int nid)
7d6034d3 3317{
3f37a36b 3318 struct pci_dev *F3 = node_to_amd_nb(nid)->misc;
0092b20d 3319 struct amd64_family_type *fam_type = NULL;
360b7f3c 3320 struct mem_ctl_info *mci = NULL;
ab5a503c 3321 struct edac_mc_layer layers[2];
3f37a36b 3322 struct amd64_pvt *pvt = NULL;
936fc3af 3323 u16 pci_id1, pci_id2;
7d6034d3
DT
3324 int err = 0, ret;
3325
3326 ret = -ENOMEM;
3327 pvt = kzalloc(sizeof(struct amd64_pvt), GFP_KERNEL);
3328 if (!pvt)
360b7f3c 3329 goto err_ret;
7d6034d3 3330
360b7f3c 3331 pvt->mc_node_id = nid;
3f37a36b 3332 pvt->F3 = F3;
7d6034d3 3333
395ae783 3334 ret = -EINVAL;
d1ea71cd 3335 fam_type = per_family_init(pvt);
0092b20d 3336 if (!fam_type)
395ae783
BP
3337 goto err_free;
3338
936fc3af 3339 if (pvt->fam >= 0x17) {
bdcee774 3340 pvt->umc = kcalloc(num_umcs, sizeof(struct amd64_umc), GFP_KERNEL);
936fc3af
YG
3341 if (!pvt->umc) {
3342 ret = -ENOMEM;
3343 goto err_free;
3344 }
3345
3346 pci_id1 = fam_type->f0_id;
3347 pci_id2 = fam_type->f6_id;
3348 } else {
3349 pci_id1 = fam_type->f1_id;
3350 pci_id2 = fam_type->f2_id;
3351 }
3352
3353 err = reserve_mc_sibling_devs(pvt, pci_id1, pci_id2);
7d6034d3 3354 if (err)
936fc3af 3355 goto err_post_init;
7d6034d3 3356
360b7f3c 3357 read_mc_regs(pvt);
7d6034d3 3358
7d6034d3
DT
3359 /*
3360 * We need to determine how many memory channels there are. Then use
3361 * that information for calculating the size of the dynamic instance
360b7f3c 3362 * tables in the 'mci' structure.
7d6034d3 3363 */
360b7f3c 3364 ret = -EINVAL;
7d6034d3
DT
3365 pvt->channel_count = pvt->ops->early_channel_count(pvt);
3366 if (pvt->channel_count < 0)
360b7f3c 3367 goto err_siblings;
7d6034d3
DT
3368
3369 ret = -ENOMEM;
ab5a503c
MCC
3370 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
3371 layers[0].size = pvt->csels[0].b_cnt;
3372 layers[0].is_virt_csrow = true;
3373 layers[1].type = EDAC_MC_LAYER_CHANNEL;
f0a56c48
BP
3374
3375 /*
3376 * Always allocate two channels since we can have setups with DIMMs on
3377 * only one channel. Also, this simplifies handling later for the price
3378 * of a couple of KBs tops.
869adc43
YG
3379 *
3380 * On Fam17h+, the number of controllers may be greater than two. So set
3381 * the size equal to the maximum number of UMCs.
f0a56c48 3382 */
869adc43
YG
3383 if (pvt->fam >= 0x17)
3384 layers[1].size = num_umcs;
3385 else
3386 layers[1].size = 2;
ab5a503c 3387 layers[1].is_virt_csrow = false;
f0a56c48 3388
ca0907b9 3389 mci = edac_mc_alloc(nid, ARRAY_SIZE(layers), layers, 0);
7d6034d3 3390 if (!mci)
360b7f3c 3391 goto err_siblings;
7d6034d3
DT
3392
3393 mci->pvt_info = pvt;
3f37a36b 3394 mci->pdev = &pvt->F3->dev;
7d6034d3 3395
df71a053 3396 setup_mci_misc_attrs(mci, fam_type);
360b7f3c
BP
3397
3398 if (init_csrows(mci))
7d6034d3
DT
3399 mci->edac_cap = EDAC_FLAG_NONE;
3400
7d6034d3 3401 ret = -ENODEV;
e339f1ec 3402 if (edac_mc_add_mc_with_groups(mci, amd64_edac_attr_groups)) {
956b9ba1 3403 edac_dbg(1, "failed edac_mc_add_mc()\n");
7d6034d3
DT
3404 goto err_add_mc;
3405 }
3406
7d6034d3
DT
3407 return 0;
3408
3409err_add_mc:
3410 edac_mc_free(mci);
3411
360b7f3c
BP
3412err_siblings:
3413 free_mc_sibling_devs(pvt);
7d6034d3 3414
936fc3af
YG
3415err_post_init:
3416 if (pvt->fam >= 0x17)
3417 kfree(pvt->umc);
3418
360b7f3c
BP
3419err_free:
3420 kfree(pvt);
7d6034d3 3421
360b7f3c 3422err_ret:
7d6034d3
DT
3423 return ret;
3424}
3425
3f37a36b 3426static int probe_one_instance(unsigned int nid)
7d6034d3 3427{
2299ef71 3428 struct pci_dev *F3 = node_to_amd_nb(nid)->misc;
ae7bb7c6 3429 struct ecc_settings *s;
3f37a36b 3430 int ret;
7d6034d3 3431
ae7bb7c6
BP
3432 ret = -ENOMEM;
3433 s = kzalloc(sizeof(struct ecc_settings), GFP_KERNEL);
3434 if (!s)
2299ef71 3435 goto err_out;
ae7bb7c6
BP
3436
3437 ecc_stngs[nid] = s;
3438
2299ef71 3439 if (!ecc_enabled(F3, nid)) {
4688c9b4 3440 ret = 0;
2299ef71
BP
3441
3442 if (!ecc_enable_override)
3443 goto err_enable;
3444
044e7a41
YG
3445 if (boot_cpu_data.x86 >= 0x17) {
3446 amd64_warn("Forcing ECC on is not recommended on newer systems. Please enable ECC in BIOS.");
3447 goto err_enable;
3448 } else
3449 amd64_warn("Forcing ECC on!\n");
2299ef71
BP
3450
3451 if (!enable_ecc_error_reporting(s, nid, F3))
3452 goto err_enable;
3453 }
3454
3f37a36b 3455 ret = init_one_instance(nid);
360b7f3c 3456 if (ret < 0) {
ae7bb7c6 3457 amd64_err("Error probing instance: %d\n", nid);
044e7a41
YG
3458
3459 if (boot_cpu_data.x86 < 0x17)
3460 restore_ecc_error_reporting(s, nid, F3);
2b9b2c46
YG
3461
3462 goto err_enable;
360b7f3c 3463 }
7d6034d3
DT
3464
3465 return ret;
2299ef71
BP
3466
3467err_enable:
3468 kfree(s);
3469 ecc_stngs[nid] = NULL;
3470
3471err_out:
3472 return ret;
7d6034d3
DT
3473}
3474
3f37a36b 3475static void remove_one_instance(unsigned int nid)
7d6034d3 3476{
360b7f3c
BP
3477 struct pci_dev *F3 = node_to_amd_nb(nid)->misc;
3478 struct ecc_settings *s = ecc_stngs[nid];
3f37a36b
BP
3479 struct mem_ctl_info *mci;
3480 struct amd64_pvt *pvt;
7d6034d3 3481
3f37a36b 3482 mci = find_mci_by_dev(&F3->dev);
a4b4bedc
BP
3483 WARN_ON(!mci);
3484
7d6034d3 3485 /* Remove from EDAC CORE tracking list */
3f37a36b 3486 mci = edac_mc_del_mc(&F3->dev);
7d6034d3
DT
3487 if (!mci)
3488 return;
3489
3490 pvt = mci->pvt_info;
3491
360b7f3c 3492 restore_ecc_error_reporting(s, nid, F3);
7d6034d3 3493
360b7f3c 3494 free_mc_sibling_devs(pvt);
7d6034d3 3495
360b7f3c
BP
3496 kfree(ecc_stngs[nid]);
3497 ecc_stngs[nid] = NULL;
ae7bb7c6 3498
7d6034d3 3499 /* Free the EDAC CORE resources */
8f68ed97 3500 mci->pvt_info = NULL;
8f68ed97
BP
3501
3502 kfree(pvt);
7d6034d3
DT
3503 edac_mc_free(mci);
3504}
3505
360b7f3c 3506static void setup_pci_device(void)
7d6034d3
DT
3507{
3508 struct mem_ctl_info *mci;
3509 struct amd64_pvt *pvt;
3510
d1ea71cd 3511 if (pci_ctl)
7d6034d3
DT
3512 return;
3513
2ec591ac 3514 mci = edac_mc_find(0);
d1ea71cd
BP
3515 if (!mci)
3516 return;
7d6034d3 3517
d1ea71cd 3518 pvt = mci->pvt_info;
936fc3af
YG
3519 if (pvt->umc)
3520 pci_ctl = edac_pci_create_generic_ctl(&pvt->F0->dev, EDAC_MOD_STR);
3521 else
3522 pci_ctl = edac_pci_create_generic_ctl(&pvt->F2->dev, EDAC_MOD_STR);
d1ea71cd
BP
3523 if (!pci_ctl) {
3524 pr_warn("%s(): Unable to create PCI control\n", __func__);
3525 pr_warn("%s(): PCI error report via EDAC not set\n", __func__);
7d6034d3
DT
3526 }
3527}
3528
d6efab74
YG
3529static const struct x86_cpu_id amd64_cpuids[] = {
3530 { X86_VENDOR_AMD, 0xF, X86_MODEL_ANY, X86_FEATURE_ANY, 0 },
3531 { X86_VENDOR_AMD, 0x10, X86_MODEL_ANY, X86_FEATURE_ANY, 0 },
3532 { X86_VENDOR_AMD, 0x15, X86_MODEL_ANY, X86_FEATURE_ANY, 0 },
3533 { X86_VENDOR_AMD, 0x16, X86_MODEL_ANY, X86_FEATURE_ANY, 0 },
95d3af6b 3534 { X86_VENDOR_AMD, 0x17, X86_MODEL_ANY, X86_FEATURE_ANY, 0 },
c4a3e946 3535 { X86_VENDOR_HYGON, 0x18, X86_MODEL_ANY, X86_FEATURE_ANY, 0 },
d6efab74
YG
3536 { }
3537};
3538MODULE_DEVICE_TABLE(x86cpu, amd64_cpuids);
3539
7d6034d3
DT
3540static int __init amd64_edac_init(void)
3541{
301375e7 3542 const char *owner;
360b7f3c 3543 int err = -ENODEV;
3f37a36b 3544 int i;
7d6034d3 3545
301375e7
TK
3546 owner = edac_get_owner();
3547 if (owner && strncmp(owner, EDAC_MOD_STR, sizeof(EDAC_MOD_STR)))
3548 return -EBUSY;
3549
1bd9900b
YG
3550 if (!x86_match_cpu(amd64_cpuids))
3551 return -ENODEV;
3552
9653a5c7 3553 if (amd_cache_northbridges() < 0)
1bd9900b 3554 return -ENODEV;
7d6034d3 3555
6ba92fea
BP
3556 opstate_init();
3557
cc4d8860 3558 err = -ENOMEM;
6396bb22 3559 ecc_stngs = kcalloc(amd_nb_num(), sizeof(ecc_stngs[0]), GFP_KERNEL);
2ec591ac 3560 if (!ecc_stngs)
a9f0fbe2 3561 goto err_free;
cc4d8860 3562
50542251 3563 msrs = msrs_alloc();
56b34b91 3564 if (!msrs)
360b7f3c 3565 goto err_free;
50542251 3566
bdcee774
YG
3567 compute_num_umcs();
3568
2287c636
YG
3569 for (i = 0; i < amd_nb_num(); i++) {
3570 err = probe_one_instance(i);
3571 if (err) {
3f37a36b
BP
3572 /* unwind properly */
3573 while (--i >= 0)
3574 remove_one_instance(i);
7d6034d3 3575
3f37a36b
BP
3576 goto err_pci;
3577 }
2287c636 3578 }
7d6034d3 3579
4688c9b4
YG
3580 if (!edac_has_mcs()) {
3581 err = -ENODEV;
3582 goto err_pci;
3583 }
3584
234365f5
YG
3585 /* register stuff with EDAC MCE */
3586 if (report_gart_errors)
3587 amd_report_gart_errors(true);
3588
3589 if (boot_cpu_data.x86 >= 0x17)
3590 amd_register_ecc_decoder(decode_umc_error);
3591 else
3592 amd_register_ecc_decoder(decode_bus_error);
3593
360b7f3c 3594 setup_pci_device();
f5b10c45
TP
3595
3596#ifdef CONFIG_X86_32
3597 amd64_err("%s on 32-bit is unsupported. USE AT YOUR OWN RISK!\n", EDAC_MOD_STR);
3598#endif
3599
de0336b3
BP
3600 printk(KERN_INFO "AMD64 EDAC driver v%s\n", EDAC_AMD64_VERSION);
3601
360b7f3c 3602 return 0;
7d6034d3 3603
56b34b91
BP
3604err_pci:
3605 msrs_free(msrs);
3606 msrs = NULL;
cc4d8860 3607
360b7f3c 3608err_free:
360b7f3c
BP
3609 kfree(ecc_stngs);
3610 ecc_stngs = NULL;
3611
7d6034d3
DT
3612 return err;
3613}
3614
3615static void __exit amd64_edac_exit(void)
3616{
3f37a36b
BP
3617 int i;
3618
d1ea71cd
BP
3619 if (pci_ctl)
3620 edac_pci_release_generic_ctl(pci_ctl);
7d6034d3 3621
234365f5
YG
3622 /* unregister from EDAC MCE */
3623 amd_report_gart_errors(false);
3624
3625 if (boot_cpu_data.x86 >= 0x17)
3626 amd_unregister_ecc_decoder(decode_umc_error);
3627 else
3628 amd_unregister_ecc_decoder(decode_bus_error);
3629
3f37a36b
BP
3630 for (i = 0; i < amd_nb_num(); i++)
3631 remove_one_instance(i);
50542251 3632
ae7bb7c6
BP
3633 kfree(ecc_stngs);
3634 ecc_stngs = NULL;
3635
50542251
BP
3636 msrs_free(msrs);
3637 msrs = NULL;
7d6034d3
DT
3638}
3639
3640module_init(amd64_edac_init);
3641module_exit(amd64_edac_exit);
3642
3643MODULE_LICENSE("GPL");
3644MODULE_AUTHOR("SoftwareBitMaker: Doug Thompson, "
3645 "Dave Peterson, Thayne Harbaugh");
3646MODULE_DESCRIPTION("MC support for AMD64 memory controllers - "
3647 EDAC_AMD64_VERSION);
3648
3649module_param(edac_op_state, int, 0444);
3650MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");