1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * cpc925_edac.c, EDAC driver for IBM CPC925 Bridge and Memory Controller.
4 *
5 * Copyright (c) 2008 Wind River Systems, Inc.
6 *
7 * Authors:	Cao Qingtao <qingtao.cao@windriver.com>
8 */
9
10#include <linux/module.h>
11#include <linux/init.h>
12#include <linux/io.h>
13#include <linux/edac.h>
14#include <linux/of.h>
15#include <linux/platform_device.h>
16#include <linux/gfp.h>
17
18#include "edac_module.h"
19
20#define CPC925_EDAC_REVISION	" Ver: 1.0.0"
21#define CPC925_EDAC_MOD_STR	"cpc925_edac"
22
23#define cpc925_printk(level, fmt, arg...) \
24	edac_printk(level, "CPC925", fmt, ##arg)
25
26#define cpc925_mc_printk(mci, level, fmt, arg...) \
27	edac_mc_chipset_printk(mci, level, "CPC925", fmt, ##arg)
28
29/*
30 * CPC925 registers are of 32 bits with bit0 defined at the
31 * most significant bit and bit31 at that of least significant.
32 */
33#define CPC925_BITS_PER_REG	32
34#define CPC925_BIT(nr)		(1UL << (CPC925_BITS_PER_REG - 1 - nr))
35
36/*
37 * EDAC device names for the error detections of
38 * CPU Interface and Hypertransport Link.
39 */
40#define CPC925_CPU_ERR_DEV	"cpu"
41#define CPC925_HT_LINK_DEV	"htlink"
42
43/* Suppose DDR Refresh cycle is 15.6 microsecond */
44#define CPC925_REF_FREQ		0xFA69
45#define CPC925_SCRUB_BLOCK_SIZE 64	/* bytes */
46#define CPC925_NR_CSROWS	8
47
48/*
49 * All registers and bits definitions are taken from
50 * "CPC925 Bridge and Memory Controller User Manual, SA14-2761-02".
51 */
52
53/*
54 * CPU and Memory Controller Registers
55 */
56/************************************************************
57 *	Processor Interface Exception Mask Register (APIMASK)
58 ************************************************************/
59#define REG_APIMASK_OFFSET	0x30070
60enum apimask_bits {
61	APIMASK_DART	= CPC925_BIT(0), /* DART Exception */
62	APIMASK_ADI0	= CPC925_BIT(1), /* Handshake Error on PI0_ADI */
63	APIMASK_ADI1	= CPC925_BIT(2), /* Handshake Error on PI1_ADI */
64	APIMASK_STAT	= CPC925_BIT(3), /* Status Exception */
65	APIMASK_DERR	= CPC925_BIT(4), /* Data Error Exception */
66	APIMASK_ADRS0	= CPC925_BIT(5), /* Addressing Exception on PI0 */
67	APIMASK_ADRS1	= CPC925_BIT(6), /* Addressing Exception on PI1 */
68					 /* BIT(7) Reserved */
69	APIMASK_ECC_UE_H = CPC925_BIT(8), /* UECC upper */
70	APIMASK_ECC_CE_H = CPC925_BIT(9), /* CECC upper */
71	APIMASK_ECC_UE_L = CPC925_BIT(10), /* UECC lower */
72	APIMASK_ECC_CE_L = CPC925_BIT(11), /* CECC lower */
73
74	CPU_MASK_ENABLE = (APIMASK_DART | APIMASK_ADI0 | APIMASK_ADI1 |
75			   APIMASK_STAT | APIMASK_DERR | APIMASK_ADRS0 |
76			   APIMASK_ADRS1),
77	ECC_MASK_ENABLE = (APIMASK_ECC_UE_H | APIMASK_ECC_CE_H |
78			   APIMASK_ECC_UE_L | APIMASK_ECC_CE_L),
79};
80#define APIMASK_ADI(n)		CPC925_BIT(((n)+1))
81
82/************************************************************
83 *	Processor Interface Exception Register (APIEXCP)
84 ************************************************************/
85#define REG_APIEXCP_OFFSET	0x30060
86enum apiexcp_bits {
87	APIEXCP_DART	= CPC925_BIT(0), /* DART Exception */
88	APIEXCP_ADI0	= CPC925_BIT(1), /* Handshake Error on PI0_ADI */
89	APIEXCP_ADI1	= CPC925_BIT(2), /* Handshake Error on PI1_ADI */
90	APIEXCP_STAT	= CPC925_BIT(3), /* Status Exception */
91	APIEXCP_DERR	= CPC925_BIT(4), /* Data Error Exception */
92	APIEXCP_ADRS0	= CPC925_BIT(5), /* Addressing Exception on PI0 */
93	APIEXCP_ADRS1	= CPC925_BIT(6), /* Addressing Exception on PI1 */
94					 /* BIT(7) Reserved */
95	APIEXCP_ECC_UE_H = CPC925_BIT(8), /* UECC upper */
96	APIEXCP_ECC_CE_H = CPC925_BIT(9), /* CECC upper */
97	APIEXCP_ECC_UE_L = CPC925_BIT(10), /* UECC lower */
98	APIEXCP_ECC_CE_L = CPC925_BIT(11), /* CECC lower */
99
100	CPU_EXCP_DETECTED = (APIEXCP_DART | APIEXCP_ADI0 | APIEXCP_ADI1 |
101			     APIEXCP_STAT | APIEXCP_DERR | APIEXCP_ADRS0 |
102			     APIEXCP_ADRS1),
103	UECC_EXCP_DETECTED = (APIEXCP_ECC_UE_H | APIEXCP_ECC_UE_L),
104	CECC_EXCP_DETECTED = (APIEXCP_ECC_CE_H | APIEXCP_ECC_CE_L),
105	ECC_EXCP_DETECTED = (UECC_EXCP_DETECTED | CECC_EXCP_DETECTED),
106};
107
108/************************************************************
109 *	Memory Bus Configuration Register (MBCR)
110************************************************************/
111#define REG_MBCR_OFFSET		0x2190
112#define MBCR_64BITCFG_SHIFT	23
113#define MBCR_64BITCFG_MASK	(1UL << MBCR_64BITCFG_SHIFT)
114#define MBCR_64BITBUS_SHIFT	22
115#define MBCR_64BITBUS_MASK	(1UL << MBCR_64BITBUS_SHIFT)
116
117/************************************************************
118 *	Memory Bank Mode Register (MBMR)
119************************************************************/
120#define REG_MBMR_OFFSET		0x21C0
121#define MBMR_MODE_MAX_VALUE	0xF
122#define MBMR_MODE_SHIFT		25
123#define MBMR_MODE_MASK		(MBMR_MODE_MAX_VALUE << MBMR_MODE_SHIFT)
124#define MBMR_BBA_SHIFT		24
125#define MBMR_BBA_MASK		(1UL << MBMR_BBA_SHIFT)
126
127/************************************************************
128 *	Memory Bank Boundary Address Register (MBBAR)
129 ************************************************************/
130#define REG_MBBAR_OFFSET	0x21D0
131#define MBBAR_BBA_MAX_VALUE	0xFF
132#define MBBAR_BBA_SHIFT		24
133#define MBBAR_BBA_MASK		(MBBAR_BBA_MAX_VALUE << MBBAR_BBA_SHIFT)
134
135/************************************************************
136 *	Memory Scrub Control Register (MSCR)
137 ************************************************************/
138#define REG_MSCR_OFFSET		0x2400
139#define MSCR_SCRUB_MOD_MASK	0xC0000000 /* scrub_mod - bit0:1*/
140#define MSCR_BACKGR_SCRUB	0x40000000 /* 01 */
141#define MSCR_SI_SHIFT		16 	/* si - bit8:15*/
142#define MSCR_SI_MAX_VALUE	0xFF
143#define MSCR_SI_MASK		(MSCR_SI_MAX_VALUE << MSCR_SI_SHIFT)
144
145/************************************************************
146 *	Memory Scrub Range Start Register (MSRSR)
147 ************************************************************/
148#define REG_MSRSR_OFFSET	0x2410
149
150/************************************************************
151 *	Memory Scrub Range End Register (MSRER)
152 ************************************************************/
153#define REG_MSRER_OFFSET	0x2420
154
155/************************************************************
156 *	Memory Scrub Pattern Register (MSPR)
157 ************************************************************/
158#define REG_MSPR_OFFSET		0x2430
159
160/************************************************************
161 *	Memory Check Control Register (MCCR)
162 ************************************************************/
163#define REG_MCCR_OFFSET		0x2440
164enum mccr_bits {
165	MCCR_ECC_EN	= CPC925_BIT(0), /* ECC high and low check */
166};
167
168/************************************************************
169 *	Memory Check Range End Register (MCRER)
170 ************************************************************/
171#define REG_MCRER_OFFSET	0x2450
172
173/************************************************************
174 *	Memory Error Address Register (MEAR)
175 ************************************************************/
176#define REG_MEAR_OFFSET		0x2460
177#define MEAR_BCNT_MAX_VALUE	0x3
178#define MEAR_BCNT_SHIFT		30
179#define MEAR_BCNT_MASK		(MEAR_BCNT_MAX_VALUE << MEAR_BCNT_SHIFT)
180#define MEAR_RANK_MAX_VALUE	0x7
181#define MEAR_RANK_SHIFT		27
182#define MEAR_RANK_MASK		(MEAR_RANK_MAX_VALUE << MEAR_RANK_SHIFT)
183#define MEAR_COL_MAX_VALUE	0x7FF
184#define MEAR_COL_SHIFT		16
185#define MEAR_COL_MASK		(MEAR_COL_MAX_VALUE << MEAR_COL_SHIFT)
186#define MEAR_BANK_MAX_VALUE	0x3
187#define MEAR_BANK_SHIFT		14
188#define MEAR_BANK_MASK		(MEAR_BANK_MAX_VALUE << MEAR_BANK_SHIFT)
189#define MEAR_ROW_MASK		0x00003FFF
190
191/************************************************************
192 *	Memory Error Syndrome Register (MESR)
193 ************************************************************/
194#define REG_MESR_OFFSET		0x2470
195#define MESR_ECC_SYN_H_MASK	0xFF00
196#define MESR_ECC_SYN_L_MASK	0x00FF
197
198/************************************************************
199 *	Memory Mode Control Register (MMCR)
200 ************************************************************/
201#define REG_MMCR_OFFSET		0x2500
202enum mmcr_bits {
203	MMCR_REG_DIMM_MODE = CPC925_BIT(3),
204};
205
206/*
207 * HyperTransport Link Registers
208 */
209/************************************************************
210 *  Error Handling/Enumeration Scratch Pad Register (ERRCTRL)
211 ************************************************************/
212#define REG_ERRCTRL_OFFSET	0x70140
213enum errctrl_bits {			 /* nonfatal interrupts for */
214	ERRCTRL_SERR_NF	= CPC925_BIT(0), /* system error */
215	ERRCTRL_CRC_NF	= CPC925_BIT(1), /* CRC error */
216	ERRCTRL_RSP_NF	= CPC925_BIT(2), /* Response error */
217	ERRCTRL_EOC_NF	= CPC925_BIT(3), /* End-Of-Chain error */
218	ERRCTRL_OVF_NF	= CPC925_BIT(4), /* Overflow error */
219	ERRCTRL_PROT_NF	= CPC925_BIT(5), /* Protocol error */
220
221	ERRCTRL_RSP_ERR	= CPC925_BIT(6), /* Response error received */
222	ERRCTRL_CHN_FAL = CPC925_BIT(7), /* Sync flooding detected */
223
224	HT_ERRCTRL_ENABLE = (ERRCTRL_SERR_NF | ERRCTRL_CRC_NF |
225			     ERRCTRL_RSP_NF | ERRCTRL_EOC_NF |
226			     ERRCTRL_OVF_NF | ERRCTRL_PROT_NF),
227	HT_ERRCTRL_DETECTED = (ERRCTRL_RSP_ERR | ERRCTRL_CHN_FAL),
228};
229
230/************************************************************
231 *  Link Configuration and Link Control Register (LINKCTRL)
232 ************************************************************/
233#define REG_LINKCTRL_OFFSET	0x70110
234enum linkctrl_bits {
235	LINKCTRL_CRC_ERR	= (CPC925_BIT(22) | CPC925_BIT(23)),
236	LINKCTRL_LINK_FAIL	= CPC925_BIT(27),
237
238	HT_LINKCTRL_DETECTED	= (LINKCTRL_CRC_ERR | LINKCTRL_LINK_FAIL),
239};
240
241/************************************************************
242 *  Link FreqCap/Error/Freq/Revision ID Register (LINKERR)
243 ************************************************************/
244#define REG_LINKERR_OFFSET	0x70120
245enum linkerr_bits {
246	LINKERR_EOC_ERR		= CPC925_BIT(17), /* End-Of-Chain error */
247	LINKERR_OVF_ERR		= CPC925_BIT(18), /* Receive Buffer Overflow */
248	LINKERR_PROT_ERR	= CPC925_BIT(19), /* Protocol error */
249
250	HT_LINKERR_DETECTED	= (LINKERR_EOC_ERR | LINKERR_OVF_ERR |
251				   LINKERR_PROT_ERR),
252};
253
254/************************************************************
255 *	Bridge Control Register (BRGCTRL)
256 ************************************************************/
257#define REG_BRGCTRL_OFFSET	0x70300
258enum brgctrl_bits {
259	BRGCTRL_DETSERR = CPC925_BIT(0), /* SERR on Secondary Bus */
260	BRGCTRL_SECBUSRESET = CPC925_BIT(9), /* Secondary Bus Reset */
261};
262
263/* Private structure for edac memory controller */
264struct cpc925_mc_pdata {
265	void __iomem *vbase;
266	unsigned long total_mem;
267	const char *name;
268	int edac_idx;
269};
270
271/* Private structure for common edac device */
272struct cpc925_dev_info {
273	void __iomem *vbase;
274	struct platform_device *pdev;
275	char *ctl_name;
276	int edac_idx;
277	struct edac_device_ctl_info *edac_dev;
278	void (*init)(struct cpc925_dev_info *dev_info);
279	void (*exit)(struct cpc925_dev_info *dev_info);
280	void (*check)(struct edac_device_ctl_info *edac_dev);
281};
282
283/* Get total memory size from Open Firmware DTB */
284static void get_total_mem(struct cpc925_mc_pdata *pdata)
285{
286	struct device_node *np = NULL;
287	const unsigned int *reg, *reg_end;
288	int len, sw, aw;
289	unsigned long start, size;
290
291	np = of_find_node_by_type(NULL, "memory");
292	if (!np)
293		return;
294
295	aw = of_n_addr_cells(np);
296	sw = of_n_size_cells(np);
297	reg = (const unsigned int *)of_get_property(np, "reg", &len);
298	reg_end = reg + len/4;
299
300	pdata->total_mem = 0;
301	do {
302		start = of_read_number(reg, aw);
303		reg += aw;
304		size = of_read_number(reg, sw);
305		reg += sw;
306		edac_dbg(1, "start 0x%lx, size 0x%lx\n", start, size);
307		pdata->total_mem += size;
308	} while (reg < reg_end);
309
310	of_node_put(np);
311	edac_dbg(0, "total_mem 0x%lx\n", pdata->total_mem);
312}
313
314static void cpc925_init_csrows(struct mem_ctl_info *mci)
315{
316	struct cpc925_mc_pdata *pdata = mci->pvt_info;
317	struct csrow_info *csrow;
318	struct dimm_info *dimm;
319	enum dev_type dtype;
320	int index, j;
321	u32 mbmr, mbbar, bba, grain;
322	unsigned long row_size, nr_pages, last_nr_pages = 0;
323
324	get_total_mem(pdata);
325
326	for (index = 0; index < mci->nr_csrows; index++) {
327		mbmr = __raw_readl(pdata->vbase + REG_MBMR_OFFSET +
328				   0x20 * index);
329		mbbar = __raw_readl(pdata->vbase + REG_MBBAR_OFFSET +
330				   0x20 + index);
331		bba = (((mbmr & MBMR_BBA_MASK) >> MBMR_BBA_SHIFT) << 8) |
332		       ((mbbar & MBBAR_BBA_MASK) >> MBBAR_BBA_SHIFT);
333
334		if (bba == 0)
335			continue; /* not populated */
336
337		csrow = mci->csrows[index];
338
339		row_size = bba * (1UL << 28);	/* 256M */
340		csrow->first_page = last_nr_pages;
341		nr_pages = row_size >> PAGE_SHIFT;
342		csrow->last_page = csrow->first_page + nr_pages - 1;
343		last_nr_pages = csrow->last_page + 1;
344
345		switch (csrow->nr_channels) {
346		case 1: /* Single channel */
347			grain = 32; /* four-beat burst of 32 bytes */
348			break;
349		case 2: /* Dual channel */
350		default:
351			grain = 64; /* four-beat burst of 64 bytes */
352			break;
353		}
354		switch ((mbmr & MBMR_MODE_MASK) >> MBMR_MODE_SHIFT) {
355		case 6: /* 0110, no way to differentiate X8 VS X16 */
356		case 5:	/* 0101 */
357		case 8: /* 1000 */
358			dtype = DEV_X16;
359			break;
360		case 7: /* 0111 */
361		case 9: /* 1001 */
362			dtype = DEV_X8;
363			break;
364		default:
365			dtype = DEV_UNKNOWN;
366		break;
367		}
368		for (j = 0; j < csrow->nr_channels; j++) {
369			dimm = csrow->channels[j]->dimm;
370			dimm->nr_pages = nr_pages / csrow->nr_channels;
371			dimm->mtype = MEM_RDDR;
372			dimm->edac_mode = EDAC_SECDED;
373			dimm->grain = grain;
374			dimm->dtype = dtype;
375		}
376	}
377}
378
379/* Enable memory controller ECC detection */
380static void cpc925_mc_init(struct mem_ctl_info *mci)
381{
382	struct cpc925_mc_pdata *pdata = mci->pvt_info;
383	u32 apimask;
384	u32 mccr;
385
386	/* Enable various ECC error exceptions */
387	apimask = __raw_readl(pdata->vbase + REG_APIMASK_OFFSET);
388	if ((apimask & ECC_MASK_ENABLE) == 0) {
389		apimask |= ECC_MASK_ENABLE;
390		__raw_writel(apimask, pdata->vbase + REG_APIMASK_OFFSET);
391	}
392
393	/* Enable ECC detection */
394	mccr = __raw_readl(pdata->vbase + REG_MCCR_OFFSET);
395	if ((mccr & MCCR_ECC_EN) == 0) {
396		mccr |= MCCR_ECC_EN;
397		__raw_writel(mccr, pdata->vbase + REG_MCCR_OFFSET);
398	}
399}
400
401/* Disable memory controller ECC detection */
402static void cpc925_mc_exit(struct mem_ctl_info *mci)
403{
404	/*
405	 * WARNING:
406	 * We are supposed to clear the ECC error detection bits,
407	 * and it will be no problem to do so. However, once they
408	 * are cleared here if we want to re-install CPC925 EDAC
409	 * module later, setting them up in cpc925_mc_init() will
410	 * trigger machine check exception.
411	 * Also, it's ok to leave ECC error detection bits enabled,
412	 * since they are reset to 1 by default or by boot loader.
413	 */
414
415	return;
416}
417
418/*
419 * Revert DDR column/row/bank addresses into page frame number and
420 * offset in page.
421 *
422 * Suppose memory mode is 0x0111(128-bit mode, identical DIMM pairs),
423 * physical address(PA) bits to column address(CA) bits mappings are:
424 * CA	0   1   2   3   4   5   6   7   8   9   10
425 * PA	59  58  57  56  55  54  53  52  51  50  49
426 *
427 * physical address(PA) bits to bank address(BA) bits mappings are:
428 * BA	0   1
429 * PA	43  44
430 *
431 * physical address(PA) bits to row address(RA) bits mappings are:
432 * RA	0   1   2   3   4   5   6   7   8   9   10   11   12
433 * PA	36  35  34  48  47  46  45  40  41  42  39   38   37
434 */
435static void cpc925_mc_get_pfn(struct mem_ctl_info *mci, u32 mear,
436		unsigned long *pfn, unsigned long *offset, int *csrow)
437{
438	u32 bcnt, rank, col, bank, row;
439	u32 c;
440	unsigned long pa;
441	int i;
442
443	bcnt = (mear & MEAR_BCNT_MASK) >> MEAR_BCNT_SHIFT;
444	rank = (mear & MEAR_RANK_MASK) >> MEAR_RANK_SHIFT;
445	col = (mear & MEAR_COL_MASK) >> MEAR_COL_SHIFT;
446	bank = (mear & MEAR_BANK_MASK) >> MEAR_BANK_SHIFT;
447	row = mear & MEAR_ROW_MASK;
448
449	*csrow = rank;
450
451#ifdef CONFIG_EDAC_DEBUG
452	if (mci->csrows[rank]->first_page == 0) {
453		cpc925_mc_printk(mci, KERN_ERR, "ECC occurs in a "
454			"non-populated csrow, broken hardware?\n");
455		return;
456	}
457#endif
458
459	/* Revert csrow number */
460	pa = mci->csrows[rank]->first_page << PAGE_SHIFT;
461
462	/* Revert column address */
463	col += bcnt;
464	for (i = 0; i < 11; i++) {
465		c = col & 0x1;
466		col >>= 1;
467		pa |= c << (14 - i);
468	}
469
470	/* Revert bank address */
471	pa |= bank << 19;
472
473	/* Revert row address, in 4 steps */
474	for (i = 0; i < 3; i++) {
475		c = row & 0x1;
476		row >>= 1;
477		pa |= c << (26 - i);
478	}
479
480	for (i = 0; i < 3; i++) {
481		c = row & 0x1;
482		row >>= 1;
483		pa |= c << (21 + i);
484	}
485
486	for (i = 0; i < 4; i++) {
487		c = row & 0x1;
488		row >>= 1;
489		pa |= c << (18 - i);
490	}
491
492	for (i = 0; i < 3; i++) {
493		c = row & 0x1;
494		row >>= 1;
495		pa |= c << (29 - i);
496	}
497
498	*offset = pa & (PAGE_SIZE - 1);
499	*pfn = pa >> PAGE_SHIFT;
500
501	edac_dbg(0, "ECC physical address 0x%lx\n", pa);
502}
503
504static int cpc925_mc_find_channel(struct mem_ctl_info *mci, u16 syndrome)
505{
506	if ((syndrome & MESR_ECC_SYN_H_MASK) == 0)
507		return 0;
508
509	if ((syndrome & MESR_ECC_SYN_L_MASK) == 0)
510		return 1;
511
512	cpc925_mc_printk(mci, KERN_INFO, "Unexpected syndrome value: 0x%x\n",
513			 syndrome);
514	return 1;
515}
516
517/* Check memory controller registers for ECC errors */
518static void cpc925_mc_check(struct mem_ctl_info *mci)
519{
520	struct cpc925_mc_pdata *pdata = mci->pvt_info;
521	u32 apiexcp;
522	u32 mear;
523	u32 mesr;
524	u16 syndrome;
525	unsigned long pfn = 0, offset = 0;
526	int csrow = 0, channel = 0;
527
528	/* APIEXCP is cleared when read */
529	apiexcp = __raw_readl(pdata->vbase + REG_APIEXCP_OFFSET);
530	if ((apiexcp & ECC_EXCP_DETECTED) == 0)
531		return;
532
533	mesr = __raw_readl(pdata->vbase + REG_MESR_OFFSET);
534	syndrome = mesr | (MESR_ECC_SYN_H_MASK | MESR_ECC_SYN_L_MASK);
535
536	mear = __raw_readl(pdata->vbase + REG_MEAR_OFFSET);
537
538	/* Revert column/row addresses into page frame number, etc */
539	cpc925_mc_get_pfn(mci, mear, &pfn, &offset, &csrow);
540
541	if (apiexcp & CECC_EXCP_DETECTED) {
542		cpc925_mc_printk(mci, KERN_INFO, "DRAM CECC Fault\n");
543		channel = cpc925_mc_find_channel(mci, syndrome);
544		edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1,
545				     pfn, offset, syndrome,
546				     csrow, channel, -1,
547				     mci->ctl_name, "");
548	}
549
550	if (apiexcp & UECC_EXCP_DETECTED) {
551		cpc925_mc_printk(mci, KERN_INFO, "DRAM UECC Fault\n");
552		edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
553				     pfn, offset, 0,
554				     csrow, -1, -1,
555				     mci->ctl_name, "");
556	}
557
558	cpc925_mc_printk(mci, KERN_INFO, "Dump registers:\n");
559	cpc925_mc_printk(mci, KERN_INFO, "APIMASK		0x%08x\n",
560		__raw_readl(pdata->vbase + REG_APIMASK_OFFSET));
561	cpc925_mc_printk(mci, KERN_INFO, "APIEXCP		0x%08x\n",
562		apiexcp);
563	cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Ctrl	0x%08x\n",
564		__raw_readl(pdata->vbase + REG_MSCR_OFFSET));
565	cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Rge Start	0x%08x\n",
566		__raw_readl(pdata->vbase + REG_MSRSR_OFFSET));
567	cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Rge End	0x%08x\n",
568		__raw_readl(pdata->vbase + REG_MSRER_OFFSET));
569	cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Pattern	0x%08x\n",
570		__raw_readl(pdata->vbase + REG_MSPR_OFFSET));
571	cpc925_mc_printk(mci, KERN_INFO, "Mem Chk Ctrl		0x%08x\n",
572		__raw_readl(pdata->vbase + REG_MCCR_OFFSET));
573	cpc925_mc_printk(mci, KERN_INFO, "Mem Chk Rge End	0x%08x\n",
574		__raw_readl(pdata->vbase + REG_MCRER_OFFSET));
575	cpc925_mc_printk(mci, KERN_INFO, "Mem Err Address	0x%08x\n",
576		mesr);
577	cpc925_mc_printk(mci, KERN_INFO, "Mem Err Syndrome	0x%08x\n",
578		syndrome);
579}
580
581/******************** CPU err device********************************/
582static u32 cpc925_cpu_mask_disabled(void)
583{
584	struct device_node *cpunode;
585	static u32 mask = 0;
586
587	/* use cached value if available */
588	if (mask != 0)
589		return mask;
590
591	mask = APIMASK_ADI0 | APIMASK_ADI1;
592
593	for_each_of_cpu_node(cpunode) {
594		const u32 *reg = of_get_property(cpunode, "reg", NULL);
595		if (reg == NULL || *reg > 2) {
596			cpc925_printk(KERN_ERR, "Bad reg value at %pOF\n", cpunode);
597			continue;
598		}
599
600		mask &= ~APIMASK_ADI(*reg);
601	}
602
603	if (mask != (APIMASK_ADI0 | APIMASK_ADI1)) {
604		/* We assume that each CPU sits on it's own PI and that
605		 * for present CPUs the reg property equals to the PI
606		 * interface id */
607		cpc925_printk(KERN_WARNING,
608				"Assuming PI id is equal to CPU MPIC id!\n");
609	}
610
611	return mask;
612}
613
614/* Enable CPU Errors detection */
615static void cpc925_cpu_init(struct cpc925_dev_info *dev_info)
616{
617	u32 apimask;
618	u32 cpumask;
619
620	apimask = __raw_readl(dev_info->vbase + REG_APIMASK_OFFSET);
621
622	cpumask = cpc925_cpu_mask_disabled();
623	if (apimask & cpumask) {
624		cpc925_printk(KERN_WARNING, "CPU(s) not present, "
625				"but enabled in APIMASK, disabling\n");
626		apimask &= ~cpumask;
627	}
628
629	if ((apimask & CPU_MASK_ENABLE) == 0)
630		apimask |= CPU_MASK_ENABLE;
631
632	__raw_writel(apimask, dev_info->vbase + REG_APIMASK_OFFSET);
633}
634
635/* Disable CPU Errors detection */
636static void cpc925_cpu_exit(struct cpc925_dev_info *dev_info)
637{
638	/*
639	 * WARNING:
640	 * We are supposed to clear the CPU error detection bits,
641	 * and it will be no problem to do so. However, once they
642	 * are cleared here if we want to re-install CPC925 EDAC
643	 * module later, setting them up in cpc925_cpu_init() will
644	 * trigger machine check exception.
645	 * Also, it's ok to leave CPU error detection bits enabled,
646	 * since they are reset to 1 by default.
647	 */
648
649	return;
650}
651
652/* Check for CPU Errors */
653static void cpc925_cpu_check(struct edac_device_ctl_info *edac_dev)
654{
655	struct cpc925_dev_info *dev_info = edac_dev->pvt_info;
656	u32 apiexcp;
657	u32 apimask;
658
659	/* APIEXCP is cleared when read */
660	apiexcp = __raw_readl(dev_info->vbase + REG_APIEXCP_OFFSET);
661	if ((apiexcp & CPU_EXCP_DETECTED) == 0)
662		return;
663
664	if ((apiexcp & ~cpc925_cpu_mask_disabled()) == 0)
665		return;
666
667	apimask = __raw_readl(dev_info->vbase + REG_APIMASK_OFFSET);
668	cpc925_printk(KERN_INFO, "Processor Interface Fault\n"
669				 "Processor Interface register dump:\n");
670	cpc925_printk(KERN_INFO, "APIMASK		0x%08x\n", apimask);
671	cpc925_printk(KERN_INFO, "APIEXCP		0x%08x\n", apiexcp);
672
673	edac_device_handle_ue(edac_dev, 0, 0, edac_dev->ctl_name);
674}
675
676/******************** HT Link err device****************************/
677/* Enable HyperTransport Link Error detection */
678static void cpc925_htlink_init(struct cpc925_dev_info *dev_info)
679{
680	u32 ht_errctrl;
681
682	ht_errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
683	if ((ht_errctrl & HT_ERRCTRL_ENABLE) == 0) {
684		ht_errctrl |= HT_ERRCTRL_ENABLE;
685		__raw_writel(ht_errctrl, dev_info->vbase + REG_ERRCTRL_OFFSET);
686	}
687}
688
689/* Disable HyperTransport Link Error detection */
690static void cpc925_htlink_exit(struct cpc925_dev_info *dev_info)
691{
692	u32 ht_errctrl;
693
694	ht_errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
695	ht_errctrl &= ~HT_ERRCTRL_ENABLE;
696	__raw_writel(ht_errctrl, dev_info->vbase + REG_ERRCTRL_OFFSET);
697}
698
699/* Check for HyperTransport Link errors */
700static void cpc925_htlink_check(struct edac_device_ctl_info *edac_dev)
701{
702	struct cpc925_dev_info *dev_info = edac_dev->pvt_info;
703	u32 brgctrl = __raw_readl(dev_info->vbase + REG_BRGCTRL_OFFSET);
704	u32 linkctrl = __raw_readl(dev_info->vbase + REG_LINKCTRL_OFFSET);
705	u32 errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
706	u32 linkerr = __raw_readl(dev_info->vbase + REG_LINKERR_OFFSET);
707
708	if (!((brgctrl & BRGCTRL_DETSERR) ||
709	      (linkctrl & HT_LINKCTRL_DETECTED) ||
710	      (errctrl & HT_ERRCTRL_DETECTED) ||
711	      (linkerr & HT_LINKERR_DETECTED)))
712		return;
713
714	cpc925_printk(KERN_INFO, "HT Link Fault\n"
715				 "HT register dump:\n");
716	cpc925_printk(KERN_INFO, "Bridge Ctrl			0x%08x\n",
717		      brgctrl);
718	cpc925_printk(KERN_INFO, "Link Config Ctrl		0x%08x\n",
719		      linkctrl);
720	cpc925_printk(KERN_INFO, "Error Enum and Ctrl		0x%08x\n",
721		      errctrl);
722	cpc925_printk(KERN_INFO, "Link Error			0x%08x\n",
723		      linkerr);
724
725	/* Clear by write 1 */
726	if (brgctrl & BRGCTRL_DETSERR)
727		__raw_writel(BRGCTRL_DETSERR,
728				dev_info->vbase + REG_BRGCTRL_OFFSET);
729
730	if (linkctrl & HT_LINKCTRL_DETECTED)
731		__raw_writel(HT_LINKCTRL_DETECTED,
732				dev_info->vbase + REG_LINKCTRL_OFFSET);
733
734	/* Initiate Secondary Bus Reset to clear the chain failure */
735	if (errctrl & ERRCTRL_CHN_FAL)
736		__raw_writel(BRGCTRL_SECBUSRESET,
737				dev_info->vbase + REG_BRGCTRL_OFFSET);
738
739	if (errctrl & ERRCTRL_RSP_ERR)
740		__raw_writel(ERRCTRL_RSP_ERR,
741				dev_info->vbase + REG_ERRCTRL_OFFSET);
742
743	if (linkerr & HT_LINKERR_DETECTED)
744		__raw_writel(HT_LINKERR_DETECTED,
745				dev_info->vbase + REG_LINKERR_OFFSET);
746
747	edac_device_handle_ce(edac_dev, 0, 0, edac_dev->ctl_name);
748}
749
750static struct cpc925_dev_info cpc925_devs[] = {
751	{
752	.ctl_name = CPC925_CPU_ERR_DEV,
753	.init = cpc925_cpu_init,
754	.exit = cpc925_cpu_exit,
755	.check = cpc925_cpu_check,
756	},
757	{
758	.ctl_name = CPC925_HT_LINK_DEV,
759	.init = cpc925_htlink_init,
760	.exit = cpc925_htlink_exit,
761	.check = cpc925_htlink_check,
762	},
763	{ }
764};
765
766/*
767 * Add CPU Err detection and HyperTransport Link Err detection
768 * as common "edac_device", they have no corresponding device
769 * nodes in the Open Firmware DTB and we have to add platform
770 * devices for them. Also, they will share the MMIO with that
771 * of memory controller.
772 */
773static void cpc925_add_edac_devices(void __iomem *vbase)
774{
775	struct cpc925_dev_info *dev_info;
776
777	if (!vbase) {
778		cpc925_printk(KERN_ERR, "MMIO not established yet\n");
779		return;
780	}
781
782	for (dev_info = &cpc925_devs[0]; dev_info->init; dev_info++) {
783		dev_info->vbase = vbase;
784		dev_info->pdev = platform_device_register_simple(
785					dev_info->ctl_name, 0, NULL, 0);
786		if (IS_ERR(dev_info->pdev)) {
787			cpc925_printk(KERN_ERR,
788				"Can't register platform device for %s\n",
789				dev_info->ctl_name);
790			continue;
791		}
792
793		/*
794		 * Don't have to allocate private structure but
795		 * make use of cpc925_devs[] instead.
796		 */
797		dev_info->edac_idx = edac_device_alloc_index();
798		dev_info->edac_dev =
799			edac_device_alloc_ctl_info(0, dev_info->ctl_name,
800				1, NULL, 0, 0, NULL, 0, dev_info->edac_idx);
801		if (!dev_info->edac_dev) {
802			cpc925_printk(KERN_ERR, "No memory for edac device\n");
803			goto err1;
804		}
805
806		dev_info->edac_dev->pvt_info = dev_info;
807		dev_info->edac_dev->dev = &dev_info->pdev->dev;
808		dev_info->edac_dev->ctl_name = dev_info->ctl_name;
809		dev_info->edac_dev->mod_name = CPC925_EDAC_MOD_STR;
810		dev_info->edac_dev->dev_name = dev_name(&dev_info->pdev->dev);
811
812		if (edac_op_state == EDAC_OPSTATE_POLL)
813			dev_info->edac_dev->edac_check = dev_info->check;
814
815		if (dev_info->init)
816			dev_info->init(dev_info);
817
818		if (edac_device_add_device(dev_info->edac_dev) > 0) {
819			cpc925_printk(KERN_ERR,
820				"Unable to add edac device for %s\n",
821				dev_info->ctl_name);
822			goto err2;
823		}
824
825		edac_dbg(0, "Successfully added edac device for %s\n",
826			 dev_info->ctl_name);
827
828		continue;
829
830err2:
831		if (dev_info->exit)
832			dev_info->exit(dev_info);
833		edac_device_free_ctl_info(dev_info->edac_dev);
834err1:
835		platform_device_unregister(dev_info->pdev);
836	}
837}
838
839/*
840 * Delete the common "edac_device" for CPU Err Detection
841 * and HyperTransport Link Err Detection
842 */
843static void cpc925_del_edac_devices(void)
844{
845	struct cpc925_dev_info *dev_info;
846
847	for (dev_info = &cpc925_devs[0]; dev_info->init; dev_info++) {
848		if (dev_info->edac_dev) {
849			edac_device_del_device(dev_info->edac_dev->dev);
850			edac_device_free_ctl_info(dev_info->edac_dev);
851			platform_device_unregister(dev_info->pdev);
852		}
853
854		if (dev_info->exit)
855			dev_info->exit(dev_info);
856
857		edac_dbg(0, "Successfully deleted edac device for %s\n",
858			 dev_info->ctl_name);
859	}
860}
861
862/* Convert current back-ground scrub rate into byte/sec bandwidth */
863static int cpc925_get_sdram_scrub_rate(struct mem_ctl_info *mci)
864{
865	struct cpc925_mc_pdata *pdata = mci->pvt_info;
866	int bw;
867	u32 mscr;
868	u8 si;
869
870	mscr = __raw_readl(pdata->vbase + REG_MSCR_OFFSET);
871	si = (mscr & MSCR_SI_MASK) >> MSCR_SI_SHIFT;
872
873	edac_dbg(0, "Mem Scrub Ctrl Register 0x%x\n", mscr);
874
875	if (((mscr & MSCR_SCRUB_MOD_MASK) != MSCR_BACKGR_SCRUB) ||
876	    (si == 0)) {
877		cpc925_mc_printk(mci, KERN_INFO, "Scrub mode not enabled\n");
878		bw = 0;
879	} else
880		bw = CPC925_SCRUB_BLOCK_SIZE * 0xFA67 / si;
881
882	return bw;
883}
884
885/* Return 0 for single channel; 1 for dual channel */
886static int cpc925_mc_get_channels(void __iomem *vbase)
887{
888	int dual = 0;
889	u32 mbcr;
890
891	mbcr = __raw_readl(vbase + REG_MBCR_OFFSET);
892
893	/*
894	 * Dual channel only when 128-bit wide physical bus
895	 * and 128-bit configuration.
896	 */
897	if (((mbcr & MBCR_64BITCFG_MASK) == 0) &&
898	    ((mbcr & MBCR_64BITBUS_MASK) == 0))
899		dual = 1;
900
901	edac_dbg(0, "%s channel\n", (dual > 0) ? "Dual" : "Single");
902
903	return dual;
904}
905
906static int cpc925_probe(struct platform_device *pdev)
907{
908	static int edac_mc_idx;
909	struct mem_ctl_info *mci;
910	struct edac_mc_layer layers[2];
911	void __iomem *vbase;
912	struct cpc925_mc_pdata *pdata;
913	struct resource *r;
914	int res = 0, nr_channels;
915
916	edac_dbg(0, "%s platform device found!\n", pdev->name);
917
918	if (!devres_open_group(&pdev->dev, cpc925_probe, GFP_KERNEL)) {
919		res = -ENOMEM;
920		goto out;
921	}
922
923	r = platform_get_resource(pdev, IORESOURCE_MEM, 0);
924	if (!r) {
925		cpc925_printk(KERN_ERR, "Unable to get resource\n");
926		res = -ENOENT;
927		goto err1;
928	}
929
930	if (!devm_request_mem_region(&pdev->dev,
931				     r->start,
932				     resource_size(r),
933				     pdev->name)) {
934		cpc925_printk(KERN_ERR, "Unable to request mem region\n");
935		res = -EBUSY;
936		goto err1;
937	}
938
939	vbase = devm_ioremap(&pdev->dev, r->start, resource_size(r));
940	if (!vbase) {
941		cpc925_printk(KERN_ERR, "Unable to ioremap device\n");
942		res = -ENOMEM;
943		goto err2;
944	}
945
946	nr_channels = cpc925_mc_get_channels(vbase) + 1;
947
948	layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
949	layers[0].size = CPC925_NR_CSROWS;
950	layers[0].is_virt_csrow = true;
951	layers[1].type = EDAC_MC_LAYER_CHANNEL;
952	layers[1].size = nr_channels;
953	layers[1].is_virt_csrow = false;
954	mci = edac_mc_alloc(edac_mc_idx, ARRAY_SIZE(layers), layers,
955			    sizeof(struct cpc925_mc_pdata));
956	if (!mci) {
957		cpc925_printk(KERN_ERR, "No memory for mem_ctl_info\n");
958		res = -ENOMEM;
959		goto err2;
960	}
961
962	pdata = mci->pvt_info;
963	pdata->vbase = vbase;
964	pdata->edac_idx = edac_mc_idx++;
965	pdata->name = pdev->name;
966
967	mci->pdev = &pdev->dev;
968	platform_set_drvdata(pdev, mci);
969	mci->dev_name = dev_name(&pdev->dev);
970	mci->mtype_cap = MEM_FLAG_RDDR | MEM_FLAG_DDR;
971	mci->edac_ctl_cap = EDAC_FLAG_NONE | EDAC_FLAG_SECDED;
972	mci->edac_cap = EDAC_FLAG_SECDED;
973	mci->mod_name = CPC925_EDAC_MOD_STR;
974	mci->ctl_name = pdev->name;
975
976	if (edac_op_state == EDAC_OPSTATE_POLL)
977		mci->edac_check = cpc925_mc_check;
978
979	mci->ctl_page_to_phys = NULL;
980	mci->scrub_mode = SCRUB_SW_SRC;
981	mci->set_sdram_scrub_rate = NULL;
982	mci->get_sdram_scrub_rate = cpc925_get_sdram_scrub_rate;
983
984	cpc925_init_csrows(mci);
985
986	/* Setup memory controller registers */
987	cpc925_mc_init(mci);
988
989	if (edac_mc_add_mc(mci) > 0) {
990		cpc925_mc_printk(mci, KERN_ERR, "Failed edac_mc_add_mc()\n");
991		goto err3;
992	}
993
994	cpc925_add_edac_devices(vbase);
995
996	/* get this far and it's successful */
997	edac_dbg(0, "success\n");
998
999	res = 0;
1000	goto out;
1001
1002err3:
1003	cpc925_mc_exit(mci);
1004	edac_mc_free(mci);
1005err2:
1006	devm_release_mem_region(&pdev->dev, r->start, resource_size(r));
1007err1:
1008	devres_release_group(&pdev->dev, cpc925_probe);
1009out:
1010	return res;
1011}
1012
1013static int cpc925_remove(struct platform_device *pdev)
1014{
1015	struct mem_ctl_info *mci = platform_get_drvdata(pdev);
1016
1017	/*
1018	 * Delete common edac devices before edac mc, because
1019	 * the former share the MMIO of the latter.
1020	 */
1021	cpc925_del_edac_devices();
1022	cpc925_mc_exit(mci);
1023
1024	edac_mc_del_mc(&pdev->dev);
1025	edac_mc_free(mci);
1026
1027	return 0;
1028}
1029
1030static struct platform_driver cpc925_edac_driver = {
1031	.probe = cpc925_probe,
1032	.remove = cpc925_remove,
1033	.driver = {
1034		   .name = "cpc925_edac",
1035	}
1036};
1037
1038static int __init cpc925_edac_init(void)
1039{
1040	int ret = 0;
1041
1042	printk(KERN_INFO "IBM CPC925 EDAC driver " CPC925_EDAC_REVISION "\n");
1043	printk(KERN_INFO "\t(c) 2008 Wind River Systems, Inc\n");
1044
1045	/* Only support POLL mode so far */
1046	edac_op_state = EDAC_OPSTATE_POLL;
1047
1048	ret = platform_driver_register(&cpc925_edac_driver);
1049	if (ret) {
1050		printk(KERN_WARNING "Failed to register %s\n",
1051			CPC925_EDAC_MOD_STR);
1052	}
1053
1054	return ret;
1055}
1056
1057static void __exit cpc925_edac_exit(void)
1058{
1059	platform_driver_unregister(&cpc925_edac_driver);
1060}
1061
1062module_init(cpc925_edac_init);
1063module_exit(cpc925_edac_exit);
1064
1065MODULE_LICENSE("GPL");
1066MODULE_AUTHOR("Cao Qingtao <qingtao.cao@windriver.com>");
1067MODULE_DESCRIPTION("IBM CPC925 Bridge and MC EDAC kernel module");
1068