1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Xilinx SDFEC
4 *
5 * Copyright (C) 2019 Xilinx, Inc.
6 *
7 * Description:
8 * This driver is developed for SDFEC16 (Soft Decision FEC 16nm)
9 * IP. It exposes a char device which supports file operations
10 * like  open(), close() and ioctl().
11 */
12
13#include <linux/miscdevice.h>
14#include <linux/io.h>
15#include <linux/interrupt.h>
16#include <linux/kernel.h>
17#include <linux/module.h>
18#include <linux/of_platform.h>
19#include <linux/poll.h>
20#include <linux/slab.h>
21#include <linux/clk.h>
22#include <linux/compat.h>
23#include <linux/highmem.h>
24
25#include <uapi/misc/xilinx_sdfec.h>
26
27#define DEV_NAME_LEN 12
28
29static DEFINE_IDA(dev_nrs);
30
31/* Xilinx SDFEC Register Map */
32/* CODE_WRI_PROTECT Register */
33#define XSDFEC_CODE_WR_PROTECT_ADDR (0x4)
34
35/* ACTIVE Register */
36#define XSDFEC_ACTIVE_ADDR (0x8)
37#define XSDFEC_IS_ACTIVITY_SET (0x1)
38
39/* AXIS_WIDTH Register */
40#define XSDFEC_AXIS_WIDTH_ADDR (0xC)
41#define XSDFEC_AXIS_DOUT_WORDS_LSB (5)
42#define XSDFEC_AXIS_DOUT_WIDTH_LSB (3)
43#define XSDFEC_AXIS_DIN_WORDS_LSB (2)
44#define XSDFEC_AXIS_DIN_WIDTH_LSB (0)
45
46/* AXIS_ENABLE Register */
47#define XSDFEC_AXIS_ENABLE_ADDR (0x10)
48#define XSDFEC_AXIS_OUT_ENABLE_MASK (0x38)
49#define XSDFEC_AXIS_IN_ENABLE_MASK (0x7)
50#define XSDFEC_AXIS_ENABLE_MASK                                                \
51	(XSDFEC_AXIS_OUT_ENABLE_MASK | XSDFEC_AXIS_IN_ENABLE_MASK)
52
53/* FEC_CODE Register */
54#define XSDFEC_FEC_CODE_ADDR (0x14)
55
56/* ORDER Register Map */
57#define XSDFEC_ORDER_ADDR (0x18)
58
59/* Interrupt Status Register */
60#define XSDFEC_ISR_ADDR (0x1C)
61/* Interrupt Status Register Bit Mask */
62#define XSDFEC_ISR_MASK (0x3F)
63
64/* Write Only - Interrupt Enable Register */
65#define XSDFEC_IER_ADDR (0x20)
66/* Write Only - Interrupt Disable Register */
67#define XSDFEC_IDR_ADDR (0x24)
68/* Read Only - Interrupt Mask Register */
69#define XSDFEC_IMR_ADDR (0x28)
70
71/* ECC Interrupt Status Register */
72#define XSDFEC_ECC_ISR_ADDR (0x2C)
73/* Single Bit Errors */
74#define XSDFEC_ECC_ISR_SBE_MASK (0x7FF)
75/* PL Initialize Single Bit Errors */
76#define XSDFEC_PL_INIT_ECC_ISR_SBE_MASK (0x3C00000)
77/* Multi Bit Errors */
78#define XSDFEC_ECC_ISR_MBE_MASK (0x3FF800)
79/* PL Initialize Multi Bit Errors */
80#define XSDFEC_PL_INIT_ECC_ISR_MBE_MASK (0x3C000000)
81/* Multi Bit Error to Event Shift */
82#define XSDFEC_ECC_ISR_MBE_TO_EVENT_SHIFT (11)
83/* PL Initialize Multi Bit Error to Event Shift */
84#define XSDFEC_PL_INIT_ECC_ISR_MBE_TO_EVENT_SHIFT (4)
85/* ECC Interrupt Status Bit Mask */
86#define XSDFEC_ECC_ISR_MASK (XSDFEC_ECC_ISR_SBE_MASK | XSDFEC_ECC_ISR_MBE_MASK)
87/* ECC Interrupt Status PL Initialize Bit Mask */
88#define XSDFEC_PL_INIT_ECC_ISR_MASK                                            \
89	(XSDFEC_PL_INIT_ECC_ISR_SBE_MASK | XSDFEC_PL_INIT_ECC_ISR_MBE_MASK)
90/* ECC Interrupt Status All Bit Mask */
91#define XSDFEC_ALL_ECC_ISR_MASK                                                \
92	(XSDFEC_ECC_ISR_MASK | XSDFEC_PL_INIT_ECC_ISR_MASK)
93/* ECC Interrupt Status Single Bit Errors Mask */
94#define XSDFEC_ALL_ECC_ISR_SBE_MASK                                            \
95	(XSDFEC_ECC_ISR_SBE_MASK | XSDFEC_PL_INIT_ECC_ISR_SBE_MASK)
96/* ECC Interrupt Status Multi Bit Errors Mask */
97#define XSDFEC_ALL_ECC_ISR_MBE_MASK                                            \
98	(XSDFEC_ECC_ISR_MBE_MASK | XSDFEC_PL_INIT_ECC_ISR_MBE_MASK)
99
100/* Write Only - ECC Interrupt Enable Register */
101#define XSDFEC_ECC_IER_ADDR (0x30)
102/* Write Only - ECC Interrupt Disable Register */
103#define XSDFEC_ECC_IDR_ADDR (0x34)
104/* Read Only - ECC Interrupt Mask Register */
105#define XSDFEC_ECC_IMR_ADDR (0x38)
106
107/* BYPASS Register */
108#define XSDFEC_BYPASS_ADDR (0x3C)
109
110/* Turbo Code Register */
111#define XSDFEC_TURBO_ADDR (0x100)
112#define XSDFEC_TURBO_SCALE_MASK (0xFFF)
113#define XSDFEC_TURBO_SCALE_BIT_POS (8)
114#define XSDFEC_TURBO_SCALE_MAX (15)
115
116/* REG0 Register */
117#define XSDFEC_LDPC_CODE_REG0_ADDR_BASE (0x2000)
118#define XSDFEC_LDPC_CODE_REG0_ADDR_HIGH (0x27F0)
119#define XSDFEC_REG0_N_MIN (4)
120#define XSDFEC_REG0_N_MAX (32768)
121#define XSDFEC_REG0_N_MUL_P (256)
122#define XSDFEC_REG0_N_LSB (0)
123#define XSDFEC_REG0_K_MIN (2)
124#define XSDFEC_REG0_K_MAX (32766)
125#define XSDFEC_REG0_K_MUL_P (256)
126#define XSDFEC_REG0_K_LSB (16)
127
128/* REG1 Register */
129#define XSDFEC_LDPC_CODE_REG1_ADDR_BASE (0x2004)
130#define XSDFEC_LDPC_CODE_REG1_ADDR_HIGH (0x27f4)
131#define XSDFEC_REG1_PSIZE_MIN (2)
132#define XSDFEC_REG1_PSIZE_MAX (512)
133#define XSDFEC_REG1_NO_PACKING_MASK (0x400)
134#define XSDFEC_REG1_NO_PACKING_LSB (10)
135#define XSDFEC_REG1_NM_MASK (0xFF800)
136#define XSDFEC_REG1_NM_LSB (11)
137#define XSDFEC_REG1_BYPASS_MASK (0x100000)
138
139/* REG2 Register */
140#define XSDFEC_LDPC_CODE_REG2_ADDR_BASE (0x2008)
141#define XSDFEC_LDPC_CODE_REG2_ADDR_HIGH (0x27f8)
142#define XSDFEC_REG2_NLAYERS_MIN (1)
143#define XSDFEC_REG2_NLAYERS_MAX (256)
144#define XSDFEC_REG2_NNMQC_MASK (0xFFE00)
145#define XSDFEC_REG2_NMQC_LSB (9)
146#define XSDFEC_REG2_NORM_TYPE_MASK (0x100000)
147#define XSDFEC_REG2_NORM_TYPE_LSB (20)
148#define XSDFEC_REG2_SPECIAL_QC_MASK (0x200000)
149#define XSDFEC_REG2_SPEICAL_QC_LSB (21)
150#define XSDFEC_REG2_NO_FINAL_PARITY_MASK (0x400000)
151#define XSDFEC_REG2_NO_FINAL_PARITY_LSB (22)
152#define XSDFEC_REG2_MAX_SCHEDULE_MASK (0x1800000)
153#define XSDFEC_REG2_MAX_SCHEDULE_LSB (23)
154
155/* REG3 Register */
156#define XSDFEC_LDPC_CODE_REG3_ADDR_BASE (0x200C)
157#define XSDFEC_LDPC_CODE_REG3_ADDR_HIGH (0x27FC)
158#define XSDFEC_REG3_LA_OFF_LSB (8)
159#define XSDFEC_REG3_QC_OFF_LSB (16)
160
161#define XSDFEC_LDPC_REG_JUMP (0x10)
162#define XSDFEC_REG_WIDTH_JUMP (4)
163
164/* The maximum number of pinned pages */
165#define MAX_NUM_PAGES ((XSDFEC_QC_TABLE_DEPTH / PAGE_SIZE) + 1)
166
167/**
168 * struct xsdfec_clks - For managing SD-FEC clocks
169 * @core_clk: Main processing clock for core
170 * @axi_clk: AXI4-Lite memory-mapped clock
171 * @din_words_clk: DIN Words AXI4-Stream Slave clock
172 * @din_clk: DIN AXI4-Stream Slave clock
173 * @dout_clk: DOUT Words AXI4-Stream Slave clock
174 * @dout_words_clk: DOUT AXI4-Stream Slave clock
175 * @ctrl_clk: Control AXI4-Stream Slave clock
176 * @status_clk: Status AXI4-Stream Slave clock
177 */
178struct xsdfec_clks {
179	struct clk *core_clk;
180	struct clk *axi_clk;
181	struct clk *din_words_clk;
182	struct clk *din_clk;
183	struct clk *dout_clk;
184	struct clk *dout_words_clk;
185	struct clk *ctrl_clk;
186	struct clk *status_clk;
187};
188
189/**
190 * struct xsdfec_dev - Driver data for SDFEC
191 * @miscdev: Misc device handle
192 * @clks: Clocks managed by the SDFEC driver
193 * @waitq: Driver wait queue
194 * @config: Configuration of the SDFEC device
195 * @dev_name: Device name
196 * @flags: spinlock flags
197 * @regs: device physical base address
198 * @dev: pointer to device struct
199 * @state: State of the SDFEC device
200 * @error_data_lock: Error counter and states spinlock
201 * @dev_id: Device ID
202 * @isr_err_count: Count of ISR errors
203 * @cecc_count: Count of Correctable ECC errors (SBE)
204 * @uecc_count: Count of Uncorrectable ECC errors (MBE)
205 * @irq: IRQ number
206 * @state_updated: indicates State updated by interrupt handler
207 * @stats_updated: indicates Stats updated by interrupt handler
208 * @intr_enabled: indicates IRQ enabled
209 *
210 * This structure contains necessary state for SDFEC driver to operate
211 */
212struct xsdfec_dev {
213	struct miscdevice miscdev;
214	struct xsdfec_clks clks;
215	wait_queue_head_t waitq;
216	struct xsdfec_config config;
217	char dev_name[DEV_NAME_LEN];
218	unsigned long flags;
219	void __iomem *regs;
220	struct device *dev;
221	enum xsdfec_state state;
222	/* Spinlock to protect state_updated and stats_updated */
223	spinlock_t error_data_lock;
224	int dev_id;
225	u32 isr_err_count;
226	u32 cecc_count;
227	u32 uecc_count;
228	int irq;
229	bool state_updated;
230	bool stats_updated;
231	bool intr_enabled;
232};
233
234static inline void xsdfec_regwrite(struct xsdfec_dev *xsdfec, u32 addr,
235				   u32 value)
236{
237	dev_dbg(xsdfec->dev, "Writing 0x%x to offset 0x%x", value, addr);
238	iowrite32(value, xsdfec->regs + addr);
239}
240
241static inline u32 xsdfec_regread(struct xsdfec_dev *xsdfec, u32 addr)
242{
243	u32 rval;
244
245	rval = ioread32(xsdfec->regs + addr);
246	dev_dbg(xsdfec->dev, "Read value = 0x%x from offset 0x%x", rval, addr);
247	return rval;
248}
249
250static void update_bool_config_from_reg(struct xsdfec_dev *xsdfec,
251					u32 reg_offset, u32 bit_num,
252					char *config_value)
253{
254	u32 reg_val;
255	u32 bit_mask = 1 << bit_num;
256
257	reg_val = xsdfec_regread(xsdfec, reg_offset);
258	*config_value = (reg_val & bit_mask) > 0;
259}
260
261static void update_config_from_hw(struct xsdfec_dev *xsdfec)
262{
263	u32 reg_value;
264	bool sdfec_started;
265
266	/* Update the Order */
267	reg_value = xsdfec_regread(xsdfec, XSDFEC_ORDER_ADDR);
268	xsdfec->config.order = reg_value;
269
270	update_bool_config_from_reg(xsdfec, XSDFEC_BYPASS_ADDR,
271				    0, /* Bit Number, maybe change to mask */
272				    &xsdfec->config.bypass);
273
274	update_bool_config_from_reg(xsdfec, XSDFEC_CODE_WR_PROTECT_ADDR,
275				    0, /* Bit Number */
276				    &xsdfec->config.code_wr_protect);
277
278	reg_value = xsdfec_regread(xsdfec, XSDFEC_IMR_ADDR);
279	xsdfec->config.irq.enable_isr = (reg_value & XSDFEC_ISR_MASK) > 0;
280
281	reg_value = xsdfec_regread(xsdfec, XSDFEC_ECC_IMR_ADDR);
282	xsdfec->config.irq.enable_ecc_isr =
283		(reg_value & XSDFEC_ECC_ISR_MASK) > 0;
284
285	reg_value = xsdfec_regread(xsdfec, XSDFEC_AXIS_ENABLE_ADDR);
286	sdfec_started = (reg_value & XSDFEC_AXIS_IN_ENABLE_MASK) > 0;
287	if (sdfec_started)
288		xsdfec->state = XSDFEC_STARTED;
289	else
290		xsdfec->state = XSDFEC_STOPPED;
291}
292
293static int xsdfec_get_status(struct xsdfec_dev *xsdfec, void __user *arg)
294{
295	struct xsdfec_status status;
296	int err;
297
298	memset(&status, 0, sizeof(status));
299	spin_lock_irqsave(&xsdfec->error_data_lock, xsdfec->flags);
300	status.state = xsdfec->state;
301	xsdfec->state_updated = false;
302	spin_unlock_irqrestore(&xsdfec->error_data_lock, xsdfec->flags);
303	status.activity = (xsdfec_regread(xsdfec, XSDFEC_ACTIVE_ADDR) &
304			   XSDFEC_IS_ACTIVITY_SET);
305
306	err = copy_to_user(arg, &status, sizeof(status));
307	if (err)
308		err = -EFAULT;
309
310	return err;
311}
312
313static int xsdfec_get_config(struct xsdfec_dev *xsdfec, void __user *arg)
314{
315	int err;
316
317	err = copy_to_user(arg, &xsdfec->config, sizeof(xsdfec->config));
318	if (err)
319		err = -EFAULT;
320
321	return err;
322}
323
324static int xsdfec_isr_enable(struct xsdfec_dev *xsdfec, bool enable)
325{
326	u32 mask_read;
327
328	if (enable) {
329		/* Enable */
330		xsdfec_regwrite(xsdfec, XSDFEC_IER_ADDR, XSDFEC_ISR_MASK);
331		mask_read = xsdfec_regread(xsdfec, XSDFEC_IMR_ADDR);
332		if (mask_read & XSDFEC_ISR_MASK) {
333			dev_dbg(xsdfec->dev,
334				"SDFEC enabling irq with IER failed");
335			return -EIO;
336		}
337	} else {
338		/* Disable */
339		xsdfec_regwrite(xsdfec, XSDFEC_IDR_ADDR, XSDFEC_ISR_MASK);
340		mask_read = xsdfec_regread(xsdfec, XSDFEC_IMR_ADDR);
341		if ((mask_read & XSDFEC_ISR_MASK) != XSDFEC_ISR_MASK) {
342			dev_dbg(xsdfec->dev,
343				"SDFEC disabling irq with IDR failed");
344			return -EIO;
345		}
346	}
347	return 0;
348}
349
350static int xsdfec_ecc_isr_enable(struct xsdfec_dev *xsdfec, bool enable)
351{
352	u32 mask_read;
353
354	if (enable) {
355		/* Enable */
356		xsdfec_regwrite(xsdfec, XSDFEC_ECC_IER_ADDR,
357				XSDFEC_ALL_ECC_ISR_MASK);
358		mask_read = xsdfec_regread(xsdfec, XSDFEC_ECC_IMR_ADDR);
359		if (mask_read & XSDFEC_ALL_ECC_ISR_MASK) {
360			dev_dbg(xsdfec->dev,
361				"SDFEC enabling ECC irq with ECC IER failed");
362			return -EIO;
363		}
364	} else {
365		/* Disable */
366		xsdfec_regwrite(xsdfec, XSDFEC_ECC_IDR_ADDR,
367				XSDFEC_ALL_ECC_ISR_MASK);
368		mask_read = xsdfec_regread(xsdfec, XSDFEC_ECC_IMR_ADDR);
369		if (!(((mask_read & XSDFEC_ALL_ECC_ISR_MASK) ==
370		       XSDFEC_ECC_ISR_MASK) ||
371		      ((mask_read & XSDFEC_ALL_ECC_ISR_MASK) ==
372		       XSDFEC_PL_INIT_ECC_ISR_MASK))) {
373			dev_dbg(xsdfec->dev,
374				"SDFEC disable ECC irq with ECC IDR failed");
375			return -EIO;
376		}
377	}
378	return 0;
379}
380
381static int xsdfec_set_irq(struct xsdfec_dev *xsdfec, void __user *arg)
382{
383	struct xsdfec_irq irq;
384	int err;
385	int isr_err;
386	int ecc_err;
387
388	err = copy_from_user(&irq, arg, sizeof(irq));
389	if (err)
390		return -EFAULT;
391
392	/* Setup tlast related IRQ */
393	isr_err = xsdfec_isr_enable(xsdfec, irq.enable_isr);
394	if (!isr_err)
395		xsdfec->config.irq.enable_isr = irq.enable_isr;
396
397	/* Setup ECC related IRQ */
398	ecc_err = xsdfec_ecc_isr_enable(xsdfec, irq.enable_ecc_isr);
399	if (!ecc_err)
400		xsdfec->config.irq.enable_ecc_isr = irq.enable_ecc_isr;
401
402	if (isr_err < 0 || ecc_err < 0)
403		err = -EIO;
404
405	return err;
406}
407
408static int xsdfec_set_turbo(struct xsdfec_dev *xsdfec, void __user *arg)
409{
410	struct xsdfec_turbo turbo;
411	int err;
412	u32 turbo_write;
413
414	err = copy_from_user(&turbo, arg, sizeof(turbo));
415	if (err)
416		return -EFAULT;
417
418	if (turbo.alg >= XSDFEC_TURBO_ALG_MAX)
419		return -EINVAL;
420
421	if (turbo.scale > XSDFEC_TURBO_SCALE_MAX)
422		return -EINVAL;
423
424	/* Check to see what device tree says about the FEC codes */
425	if (xsdfec->config.code == XSDFEC_LDPC_CODE)
426		return -EIO;
427
428	turbo_write = ((turbo.scale & XSDFEC_TURBO_SCALE_MASK)
429		       << XSDFEC_TURBO_SCALE_BIT_POS) |
430		      turbo.alg;
431	xsdfec_regwrite(xsdfec, XSDFEC_TURBO_ADDR, turbo_write);
432	return err;
433}
434
435static int xsdfec_get_turbo(struct xsdfec_dev *xsdfec, void __user *arg)
436{
437	u32 reg_value;
438	struct xsdfec_turbo turbo_params;
439	int err;
440
441	if (xsdfec->config.code == XSDFEC_LDPC_CODE)
442		return -EIO;
443
444	memset(&turbo_params, 0, sizeof(turbo_params));
445	reg_value = xsdfec_regread(xsdfec, XSDFEC_TURBO_ADDR);
446
447	turbo_params.scale = (reg_value & XSDFEC_TURBO_SCALE_MASK) >>
448			     XSDFEC_TURBO_SCALE_BIT_POS;
449	turbo_params.alg = reg_value & 0x1;
450
451	err = copy_to_user(arg, &turbo_params, sizeof(turbo_params));
452	if (err)
453		err = -EFAULT;
454
455	return err;
456}
457
458static int xsdfec_reg0_write(struct xsdfec_dev *xsdfec, u32 n, u32 k, u32 psize,
459			     u32 offset)
460{
461	u32 wdata;
462
463	if (n < XSDFEC_REG0_N_MIN || n > XSDFEC_REG0_N_MAX || psize == 0 ||
464	    (n > XSDFEC_REG0_N_MUL_P * psize) || n <= k || ((n % psize) != 0)) {
465		dev_dbg(xsdfec->dev, "N value is not in range");
466		return -EINVAL;
467	}
468	n <<= XSDFEC_REG0_N_LSB;
469
470	if (k < XSDFEC_REG0_K_MIN || k > XSDFEC_REG0_K_MAX ||
471	    (k > XSDFEC_REG0_K_MUL_P * psize) || ((k % psize) != 0)) {
472		dev_dbg(xsdfec->dev, "K value is not in range");
473		return -EINVAL;
474	}
475	k = k << XSDFEC_REG0_K_LSB;
476	wdata = k | n;
477
478	if (XSDFEC_LDPC_CODE_REG0_ADDR_BASE + (offset * XSDFEC_LDPC_REG_JUMP) >
479	    XSDFEC_LDPC_CODE_REG0_ADDR_HIGH) {
480		dev_dbg(xsdfec->dev, "Writing outside of LDPC reg0 space 0x%x",
481			XSDFEC_LDPC_CODE_REG0_ADDR_BASE +
482				(offset * XSDFEC_LDPC_REG_JUMP));
483		return -EINVAL;
484	}
485	xsdfec_regwrite(xsdfec,
486			XSDFEC_LDPC_CODE_REG0_ADDR_BASE +
487				(offset * XSDFEC_LDPC_REG_JUMP),
488			wdata);
489	return 0;
490}
491
492static int xsdfec_reg1_write(struct xsdfec_dev *xsdfec, u32 psize,
493			     u32 no_packing, u32 nm, u32 offset)
494{
495	u32 wdata;
496
497	if (psize < XSDFEC_REG1_PSIZE_MIN || psize > XSDFEC_REG1_PSIZE_MAX) {
498		dev_dbg(xsdfec->dev, "Psize is not in range");
499		return -EINVAL;
500	}
501
502	if (no_packing != 0 && no_packing != 1)
503		dev_dbg(xsdfec->dev, "No-packing bit register invalid");
504	no_packing = ((no_packing << XSDFEC_REG1_NO_PACKING_LSB) &
505		      XSDFEC_REG1_NO_PACKING_MASK);
506
507	if (nm & ~(XSDFEC_REG1_NM_MASK >> XSDFEC_REG1_NM_LSB))
508		dev_dbg(xsdfec->dev, "NM is beyond 10 bits");
509	nm = (nm << XSDFEC_REG1_NM_LSB) & XSDFEC_REG1_NM_MASK;
510
511	wdata = nm | no_packing | psize;
512	if (XSDFEC_LDPC_CODE_REG1_ADDR_BASE + (offset * XSDFEC_LDPC_REG_JUMP) >
513	    XSDFEC_LDPC_CODE_REG1_ADDR_HIGH) {
514		dev_dbg(xsdfec->dev, "Writing outside of LDPC reg1 space 0x%x",
515			XSDFEC_LDPC_CODE_REG1_ADDR_BASE +
516				(offset * XSDFEC_LDPC_REG_JUMP));
517		return -EINVAL;
518	}
519	xsdfec_regwrite(xsdfec,
520			XSDFEC_LDPC_CODE_REG1_ADDR_BASE +
521				(offset * XSDFEC_LDPC_REG_JUMP),
522			wdata);
523	return 0;
524}
525
526static int xsdfec_reg2_write(struct xsdfec_dev *xsdfec, u32 nlayers, u32 nmqc,
527			     u32 norm_type, u32 special_qc, u32 no_final_parity,
528			     u32 max_schedule, u32 offset)
529{
530	u32 wdata;
531
532	if (nlayers < XSDFEC_REG2_NLAYERS_MIN ||
533	    nlayers > XSDFEC_REG2_NLAYERS_MAX) {
534		dev_dbg(xsdfec->dev, "Nlayers is not in range");
535		return -EINVAL;
536	}
537
538	if (nmqc & ~(XSDFEC_REG2_NNMQC_MASK >> XSDFEC_REG2_NMQC_LSB))
539		dev_dbg(xsdfec->dev, "NMQC exceeds 11 bits");
540	nmqc = (nmqc << XSDFEC_REG2_NMQC_LSB) & XSDFEC_REG2_NNMQC_MASK;
541
542	if (norm_type > 1)
543		dev_dbg(xsdfec->dev, "Norm type is invalid");
544	norm_type = ((norm_type << XSDFEC_REG2_NORM_TYPE_LSB) &
545		     XSDFEC_REG2_NORM_TYPE_MASK);
546	if (special_qc > 1)
547		dev_dbg(xsdfec->dev, "Special QC in invalid");
548	special_qc = ((special_qc << XSDFEC_REG2_SPEICAL_QC_LSB) &
549		      XSDFEC_REG2_SPECIAL_QC_MASK);
550
551	if (no_final_parity > 1)
552		dev_dbg(xsdfec->dev, "No final parity check invalid");
553	no_final_parity =
554		((no_final_parity << XSDFEC_REG2_NO_FINAL_PARITY_LSB) &
555		 XSDFEC_REG2_NO_FINAL_PARITY_MASK);
556	if (max_schedule &
557	    ~(XSDFEC_REG2_MAX_SCHEDULE_MASK >> XSDFEC_REG2_MAX_SCHEDULE_LSB))
558		dev_dbg(xsdfec->dev, "Max Schedule exceeds 2 bits");
559	max_schedule = ((max_schedule << XSDFEC_REG2_MAX_SCHEDULE_LSB) &
560			XSDFEC_REG2_MAX_SCHEDULE_MASK);
561
562	wdata = (max_schedule | no_final_parity | special_qc | norm_type |
563		 nmqc | nlayers);
564
565	if (XSDFEC_LDPC_CODE_REG2_ADDR_BASE + (offset * XSDFEC_LDPC_REG_JUMP) >
566	    XSDFEC_LDPC_CODE_REG2_ADDR_HIGH) {
567		dev_dbg(xsdfec->dev, "Writing outside of LDPC reg2 space 0x%x",
568			XSDFEC_LDPC_CODE_REG2_ADDR_BASE +
569				(offset * XSDFEC_LDPC_REG_JUMP));
570		return -EINVAL;
571	}
572	xsdfec_regwrite(xsdfec,
573			XSDFEC_LDPC_CODE_REG2_ADDR_BASE +
574				(offset * XSDFEC_LDPC_REG_JUMP),
575			wdata);
576	return 0;
577}
578
579static int xsdfec_reg3_write(struct xsdfec_dev *xsdfec, u8 sc_off, u8 la_off,
580			     u16 qc_off, u32 offset)
581{
582	u32 wdata;
583
584	wdata = ((qc_off << XSDFEC_REG3_QC_OFF_LSB) |
585		 (la_off << XSDFEC_REG3_LA_OFF_LSB) | sc_off);
586	if (XSDFEC_LDPC_CODE_REG3_ADDR_BASE + (offset * XSDFEC_LDPC_REG_JUMP) >
587	    XSDFEC_LDPC_CODE_REG3_ADDR_HIGH) {
588		dev_dbg(xsdfec->dev, "Writing outside of LDPC reg3 space 0x%x",
589			XSDFEC_LDPC_CODE_REG3_ADDR_BASE +
590				(offset * XSDFEC_LDPC_REG_JUMP));
591		return -EINVAL;
592	}
593	xsdfec_regwrite(xsdfec,
594			XSDFEC_LDPC_CODE_REG3_ADDR_BASE +
595				(offset * XSDFEC_LDPC_REG_JUMP),
596			wdata);
597	return 0;
598}
599
600static int xsdfec_table_write(struct xsdfec_dev *xsdfec, u32 offset,
601			      u32 *src_ptr, u32 len, const u32 base_addr,
602			      const u32 depth)
603{
604	u32 reg = 0;
605	int res, i, nr_pages;
606	u32 n;
607	u32 *addr = NULL;
608	struct page *pages[MAX_NUM_PAGES];
609
610	/*
611	 * Writes that go beyond the length of
612	 * Shared Scale(SC) table should fail
613	 */
614	if (offset > depth / XSDFEC_REG_WIDTH_JUMP ||
615	    len > depth / XSDFEC_REG_WIDTH_JUMP ||
616	    offset + len > depth / XSDFEC_REG_WIDTH_JUMP) {
617		dev_dbg(xsdfec->dev, "Write exceeds SC table length");
618		return -EINVAL;
619	}
620
621	n = (len * XSDFEC_REG_WIDTH_JUMP) / PAGE_SIZE;
622	if ((len * XSDFEC_REG_WIDTH_JUMP) % PAGE_SIZE)
623		n += 1;
624
625	if (WARN_ON_ONCE(n > INT_MAX))
626		return -EINVAL;
627
628	nr_pages = n;
629
630	res = pin_user_pages_fast((unsigned long)src_ptr, nr_pages, 0, pages);
631	if (res < nr_pages) {
632		if (res > 0)
633			unpin_user_pages(pages, res);
634
635		return -EINVAL;
636	}
637
638	for (i = 0; i < nr_pages; i++) {
639		addr = kmap(pages[i]);
640		do {
641			xsdfec_regwrite(xsdfec,
642					base_addr + ((offset + reg) *
643						     XSDFEC_REG_WIDTH_JUMP),
644					addr[reg]);
645			reg++;
646		} while ((reg < len) &&
647			 ((reg * XSDFEC_REG_WIDTH_JUMP) % PAGE_SIZE));
648		unpin_user_page(pages[i]);
649	}
650	return 0;
651}
652
653static int xsdfec_add_ldpc(struct xsdfec_dev *xsdfec, void __user *arg)
654{
655	struct xsdfec_ldpc_params *ldpc;
656	int ret, n;
657
658	ldpc = memdup_user(arg, sizeof(*ldpc));
659	if (IS_ERR(ldpc))
660		return PTR_ERR(ldpc);
661
662	if (xsdfec->config.code == XSDFEC_TURBO_CODE) {
663		ret = -EIO;
664		goto err_out;
665	}
666
667	/* Verify Device has not started */
668	if (xsdfec->state == XSDFEC_STARTED) {
669		ret = -EIO;
670		goto err_out;
671	}
672
673	if (xsdfec->config.code_wr_protect) {
674		ret = -EIO;
675		goto err_out;
676	}
677
678	/* Write Reg 0 */
679	ret = xsdfec_reg0_write(xsdfec, ldpc->n, ldpc->k, ldpc->psize,
680				ldpc->code_id);
681	if (ret)
682		goto err_out;
683
684	/* Write Reg 1 */
685	ret = xsdfec_reg1_write(xsdfec, ldpc->psize, ldpc->no_packing, ldpc->nm,
686				ldpc->code_id);
687	if (ret)
688		goto err_out;
689
690	/* Write Reg 2 */
691	ret = xsdfec_reg2_write(xsdfec, ldpc->nlayers, ldpc->nmqc,
692				ldpc->norm_type, ldpc->special_qc,
693				ldpc->no_final_parity, ldpc->max_schedule,
694				ldpc->code_id);
695	if (ret)
696		goto err_out;
697
698	/* Write Reg 3 */
699	ret = xsdfec_reg3_write(xsdfec, ldpc->sc_off, ldpc->la_off,
700				ldpc->qc_off, ldpc->code_id);
701	if (ret)
702		goto err_out;
703
704	/* Write Shared Codes */
705	n = ldpc->nlayers / 4;
706	if (ldpc->nlayers % 4)
707		n++;
708
709	ret = xsdfec_table_write(xsdfec, ldpc->sc_off, ldpc->sc_table, n,
710				 XSDFEC_LDPC_SC_TABLE_ADDR_BASE,
711				 XSDFEC_SC_TABLE_DEPTH);
712	if (ret < 0)
713		goto err_out;
714
715	ret = xsdfec_table_write(xsdfec, 4 * ldpc->la_off, ldpc->la_table,
716				 ldpc->nlayers, XSDFEC_LDPC_LA_TABLE_ADDR_BASE,
717				 XSDFEC_LA_TABLE_DEPTH);
718	if (ret < 0)
719		goto err_out;
720
721	ret = xsdfec_table_write(xsdfec, 4 * ldpc->qc_off, ldpc->qc_table,
722				 ldpc->nqc, XSDFEC_LDPC_QC_TABLE_ADDR_BASE,
723				 XSDFEC_QC_TABLE_DEPTH);
724err_out:
725	kfree(ldpc);
726	return ret;
727}
728
729static int xsdfec_set_order(struct xsdfec_dev *xsdfec, void __user *arg)
730{
731	bool order_invalid;
732	enum xsdfec_order order;
733	int err;
734
735	err = get_user(order, (enum xsdfec_order __user *)arg);
736	if (err)
737		return -EFAULT;
738
739	order_invalid = (order != XSDFEC_MAINTAIN_ORDER) &&
740			(order != XSDFEC_OUT_OF_ORDER);
741	if (order_invalid)
742		return -EINVAL;
743
744	/* Verify Device has not started */
745	if (xsdfec->state == XSDFEC_STARTED)
746		return -EIO;
747
748	xsdfec_regwrite(xsdfec, XSDFEC_ORDER_ADDR, order);
749
750	xsdfec->config.order = order;
751
752	return 0;
753}
754
755static int xsdfec_set_bypass(struct xsdfec_dev *xsdfec, bool __user *arg)
756{
757	bool bypass;
758	int err;
759
760	err = get_user(bypass, arg);
761	if (err)
762		return -EFAULT;
763
764	/* Verify Device has not started */
765	if (xsdfec->state == XSDFEC_STARTED)
766		return -EIO;
767
768	if (bypass)
769		xsdfec_regwrite(xsdfec, XSDFEC_BYPASS_ADDR, 1);
770	else
771		xsdfec_regwrite(xsdfec, XSDFEC_BYPASS_ADDR, 0);
772
773	xsdfec->config.bypass = bypass;
774
775	return 0;
776}
777
778static int xsdfec_is_active(struct xsdfec_dev *xsdfec, bool __user *arg)
779{
780	u32 reg_value;
781	bool is_active;
782	int err;
783
784	reg_value = xsdfec_regread(xsdfec, XSDFEC_ACTIVE_ADDR);
785	/* using a double ! operator instead of casting */
786	is_active = !!(reg_value & XSDFEC_IS_ACTIVITY_SET);
787	err = put_user(is_active, arg);
788	if (err)
789		return -EFAULT;
790
791	return err;
792}
793
794static u32
795xsdfec_translate_axis_width_cfg_val(enum xsdfec_axis_width axis_width_cfg)
796{
797	u32 axis_width_field = 0;
798
799	switch (axis_width_cfg) {
800	case XSDFEC_1x128b:
801		axis_width_field = 0;
802		break;
803	case XSDFEC_2x128b:
804		axis_width_field = 1;
805		break;
806	case XSDFEC_4x128b:
807		axis_width_field = 2;
808		break;
809	}
810
811	return axis_width_field;
812}
813
814static u32 xsdfec_translate_axis_words_cfg_val(enum xsdfec_axis_word_include
815	axis_word_inc_cfg)
816{
817	u32 axis_words_field = 0;
818
819	if (axis_word_inc_cfg == XSDFEC_FIXED_VALUE ||
820	    axis_word_inc_cfg == XSDFEC_IN_BLOCK)
821		axis_words_field = 0;
822	else if (axis_word_inc_cfg == XSDFEC_PER_AXI_TRANSACTION)
823		axis_words_field = 1;
824
825	return axis_words_field;
826}
827
828static int xsdfec_cfg_axi_streams(struct xsdfec_dev *xsdfec)
829{
830	u32 reg_value;
831	u32 dout_words_field;
832	u32 dout_width_field;
833	u32 din_words_field;
834	u32 din_width_field;
835	struct xsdfec_config *config = &xsdfec->config;
836
837	/* translate config info to register values */
838	dout_words_field =
839		xsdfec_translate_axis_words_cfg_val(config->dout_word_include);
840	dout_width_field =
841		xsdfec_translate_axis_width_cfg_val(config->dout_width);
842	din_words_field =
843		xsdfec_translate_axis_words_cfg_val(config->din_word_include);
844	din_width_field =
845		xsdfec_translate_axis_width_cfg_val(config->din_width);
846
847	reg_value = dout_words_field << XSDFEC_AXIS_DOUT_WORDS_LSB;
848	reg_value |= dout_width_field << XSDFEC_AXIS_DOUT_WIDTH_LSB;
849	reg_value |= din_words_field << XSDFEC_AXIS_DIN_WORDS_LSB;
850	reg_value |= din_width_field << XSDFEC_AXIS_DIN_WIDTH_LSB;
851
852	xsdfec_regwrite(xsdfec, XSDFEC_AXIS_WIDTH_ADDR, reg_value);
853
854	return 0;
855}
856
857static int xsdfec_dev_open(struct inode *iptr, struct file *fptr)
858{
859	return 0;
860}
861
862static int xsdfec_dev_release(struct inode *iptr, struct file *fptr)
863{
864	return 0;
865}
866
867static int xsdfec_start(struct xsdfec_dev *xsdfec)
868{
869	u32 regread;
870
871	regread = xsdfec_regread(xsdfec, XSDFEC_FEC_CODE_ADDR);
872	regread &= 0x1;
873	if (regread != xsdfec->config.code) {
874		dev_dbg(xsdfec->dev,
875			"%s SDFEC HW code does not match driver code, reg %d, code %d",
876			__func__, regread, xsdfec->config.code);
877		return -EINVAL;
878	}
879
880	/* Set AXIS enable */
881	xsdfec_regwrite(xsdfec, XSDFEC_AXIS_ENABLE_ADDR,
882			XSDFEC_AXIS_ENABLE_MASK);
883	/* Done */
884	xsdfec->state = XSDFEC_STARTED;
885	return 0;
886}
887
888static int xsdfec_stop(struct xsdfec_dev *xsdfec)
889{
890	u32 regread;
891
892	if (xsdfec->state != XSDFEC_STARTED)
893		dev_dbg(xsdfec->dev, "Device not started correctly");
894	/* Disable AXIS_ENABLE Input interfaces only */
895	regread = xsdfec_regread(xsdfec, XSDFEC_AXIS_ENABLE_ADDR);
896	regread &= (~XSDFEC_AXIS_IN_ENABLE_MASK);
897	xsdfec_regwrite(xsdfec, XSDFEC_AXIS_ENABLE_ADDR, regread);
898	/* Stop */
899	xsdfec->state = XSDFEC_STOPPED;
900	return 0;
901}
902
903static int xsdfec_clear_stats(struct xsdfec_dev *xsdfec)
904{
905	spin_lock_irqsave(&xsdfec->error_data_lock, xsdfec->flags);
906	xsdfec->isr_err_count = 0;
907	xsdfec->uecc_count = 0;
908	xsdfec->cecc_count = 0;
909	spin_unlock_irqrestore(&xsdfec->error_data_lock, xsdfec->flags);
910
911	return 0;
912}
913
914static int xsdfec_get_stats(struct xsdfec_dev *xsdfec, void __user *arg)
915{
916	int err;
917	struct xsdfec_stats user_stats;
918
919	spin_lock_irqsave(&xsdfec->error_data_lock, xsdfec->flags);
920	user_stats.isr_err_count = xsdfec->isr_err_count;
921	user_stats.cecc_count = xsdfec->cecc_count;
922	user_stats.uecc_count = xsdfec->uecc_count;
923	xsdfec->stats_updated = false;
924	spin_unlock_irqrestore(&xsdfec->error_data_lock, xsdfec->flags);
925
926	err = copy_to_user(arg, &user_stats, sizeof(user_stats));
927	if (err)
928		err = -EFAULT;
929
930	return err;
931}
932
933static int xsdfec_set_default_config(struct xsdfec_dev *xsdfec)
934{
935	/* Ensure registers are aligned with core configuration */
936	xsdfec_regwrite(xsdfec, XSDFEC_FEC_CODE_ADDR, xsdfec->config.code);
937	xsdfec_cfg_axi_streams(xsdfec);
938	update_config_from_hw(xsdfec);
939
940	return 0;
941}
942
943static long xsdfec_dev_ioctl(struct file *fptr, unsigned int cmd,
944			     unsigned long data)
945{
946	struct xsdfec_dev *xsdfec;
947	void __user *arg = NULL;
948	int rval = -EINVAL;
949
950	xsdfec = container_of(fptr->private_data, struct xsdfec_dev, miscdev);
951
952	/* In failed state allow only reset and get status IOCTLs */
953	if (xsdfec->state == XSDFEC_NEEDS_RESET &&
954	    (cmd != XSDFEC_SET_DEFAULT_CONFIG && cmd != XSDFEC_GET_STATUS &&
955	     cmd != XSDFEC_GET_STATS && cmd != XSDFEC_CLEAR_STATS)) {
956		return -EPERM;
957	}
958
959	if (_IOC_TYPE(cmd) != XSDFEC_MAGIC)
960		return -ENOTTY;
961
962	/* check if ioctl argument is present and valid */
963	if (_IOC_DIR(cmd) != _IOC_NONE) {
964		arg = (void __user *)data;
965		if (!arg)
966			return rval;
967	}
968
969	switch (cmd) {
970	case XSDFEC_START_DEV:
971		rval = xsdfec_start(xsdfec);
972		break;
973	case XSDFEC_STOP_DEV:
974		rval = xsdfec_stop(xsdfec);
975		break;
976	case XSDFEC_CLEAR_STATS:
977		rval = xsdfec_clear_stats(xsdfec);
978		break;
979	case XSDFEC_GET_STATS:
980		rval = xsdfec_get_stats(xsdfec, arg);
981		break;
982	case XSDFEC_GET_STATUS:
983		rval = xsdfec_get_status(xsdfec, arg);
984		break;
985	case XSDFEC_GET_CONFIG:
986		rval = xsdfec_get_config(xsdfec, arg);
987		break;
988	case XSDFEC_SET_DEFAULT_CONFIG:
989		rval = xsdfec_set_default_config(xsdfec);
990		break;
991	case XSDFEC_SET_IRQ:
992		rval = xsdfec_set_irq(xsdfec, arg);
993		break;
994	case XSDFEC_SET_TURBO:
995		rval = xsdfec_set_turbo(xsdfec, arg);
996		break;
997	case XSDFEC_GET_TURBO:
998		rval = xsdfec_get_turbo(xsdfec, arg);
999		break;
1000	case XSDFEC_ADD_LDPC_CODE_PARAMS:
1001		rval = xsdfec_add_ldpc(xsdfec, arg);
1002		break;
1003	case XSDFEC_SET_ORDER:
1004		rval = xsdfec_set_order(xsdfec, arg);
1005		break;
1006	case XSDFEC_SET_BYPASS:
1007		rval = xsdfec_set_bypass(xsdfec, arg);
1008		break;
1009	case XSDFEC_IS_ACTIVE:
1010		rval = xsdfec_is_active(xsdfec, (bool __user *)arg);
1011		break;
1012	default:
1013		/* Should not get here */
1014		break;
1015	}
1016	return rval;
1017}
1018
1019#ifdef CONFIG_COMPAT
1020static long xsdfec_dev_compat_ioctl(struct file *file, unsigned int cmd,
1021				    unsigned long data)
1022{
1023	return xsdfec_dev_ioctl(file, cmd, (unsigned long)compat_ptr(data));
1024}
1025#endif
1026
1027static __poll_t xsdfec_poll(struct file *file, poll_table *wait)
1028{
1029	__poll_t mask = 0;
1030	struct xsdfec_dev *xsdfec;
1031
1032	xsdfec = container_of(file->private_data, struct xsdfec_dev, miscdev);
1033
1034	if (!xsdfec)
1035		return EPOLLNVAL | EPOLLHUP;
1036
1037	poll_wait(file, &xsdfec->waitq, wait);
1038
1039	/* XSDFEC ISR detected an error */
1040	spin_lock_irqsave(&xsdfec->error_data_lock, xsdfec->flags);
1041	if (xsdfec->state_updated)
1042		mask |= EPOLLIN | EPOLLPRI;
1043
1044	if (xsdfec->stats_updated)
1045		mask |= EPOLLIN | EPOLLRDNORM;
1046	spin_unlock_irqrestore(&xsdfec->error_data_lock, xsdfec->flags);
1047
1048	return mask;
1049}
1050
1051static const struct file_operations xsdfec_fops = {
1052	.owner = THIS_MODULE,
1053	.open = xsdfec_dev_open,
1054	.release = xsdfec_dev_release,
1055	.unlocked_ioctl = xsdfec_dev_ioctl,
1056	.poll = xsdfec_poll,
1057#ifdef CONFIG_COMPAT
1058	.compat_ioctl = xsdfec_dev_compat_ioctl,
1059#endif
1060};
1061
1062static int xsdfec_parse_of(struct xsdfec_dev *xsdfec)
1063{
1064	struct device *dev = xsdfec->dev;
1065	struct device_node *node = dev->of_node;
1066	int rval;
1067	const char *fec_code;
1068	u32 din_width;
1069	u32 din_word_include;
1070	u32 dout_width;
1071	u32 dout_word_include;
1072
1073	rval = of_property_read_string(node, "xlnx,sdfec-code", &fec_code);
1074	if (rval < 0)
1075		return rval;
1076
1077	if (!strcasecmp(fec_code, "ldpc"))
1078		xsdfec->config.code = XSDFEC_LDPC_CODE;
1079	else if (!strcasecmp(fec_code, "turbo"))
1080		xsdfec->config.code = XSDFEC_TURBO_CODE;
1081	else
1082		return -EINVAL;
1083
1084	rval = of_property_read_u32(node, "xlnx,sdfec-din-words",
1085				    &din_word_include);
1086	if (rval < 0)
1087		return rval;
1088
1089	if (din_word_include < XSDFEC_AXIS_WORDS_INCLUDE_MAX)
1090		xsdfec->config.din_word_include = din_word_include;
1091	else
1092		return -EINVAL;
1093
1094	rval = of_property_read_u32(node, "xlnx,sdfec-din-width", &din_width);
1095	if (rval < 0)
1096		return rval;
1097
1098	switch (din_width) {
1099	/* Fall through and set for valid values */
1100	case XSDFEC_1x128b:
1101	case XSDFEC_2x128b:
1102	case XSDFEC_4x128b:
1103		xsdfec->config.din_width = din_width;
1104		break;
1105	default:
1106		return -EINVAL;
1107	}
1108
1109	rval = of_property_read_u32(node, "xlnx,sdfec-dout-words",
1110				    &dout_word_include);
1111	if (rval < 0)
1112		return rval;
1113
1114	if (dout_word_include < XSDFEC_AXIS_WORDS_INCLUDE_MAX)
1115		xsdfec->config.dout_word_include = dout_word_include;
1116	else
1117		return -EINVAL;
1118
1119	rval = of_property_read_u32(node, "xlnx,sdfec-dout-width", &dout_width);
1120	if (rval < 0)
1121		return rval;
1122
1123	switch (dout_width) {
1124	/* Fall through and set for valid values */
1125	case XSDFEC_1x128b:
1126	case XSDFEC_2x128b:
1127	case XSDFEC_4x128b:
1128		xsdfec->config.dout_width = dout_width;
1129		break;
1130	default:
1131		return -EINVAL;
1132	}
1133
1134	/* Write LDPC to CODE Register */
1135	xsdfec_regwrite(xsdfec, XSDFEC_FEC_CODE_ADDR, xsdfec->config.code);
1136
1137	xsdfec_cfg_axi_streams(xsdfec);
1138
1139	return 0;
1140}
1141
1142static irqreturn_t xsdfec_irq_thread(int irq, void *dev_id)
1143{
1144	struct xsdfec_dev *xsdfec = dev_id;
1145	irqreturn_t ret = IRQ_HANDLED;
1146	u32 ecc_err;
1147	u32 isr_err;
1148	u32 uecc_count;
1149	u32 cecc_count;
1150	u32 isr_err_count;
1151	u32 aecc_count;
1152	u32 tmp;
1153
1154	WARN_ON(xsdfec->irq != irq);
1155
1156	/* Mask Interrupts */
1157	xsdfec_isr_enable(xsdfec, false);
1158	xsdfec_ecc_isr_enable(xsdfec, false);
1159	/* Read ISR */
1160	ecc_err = xsdfec_regread(xsdfec, XSDFEC_ECC_ISR_ADDR);
1161	isr_err = xsdfec_regread(xsdfec, XSDFEC_ISR_ADDR);
1162	/* Clear the interrupts */
1163	xsdfec_regwrite(xsdfec, XSDFEC_ECC_ISR_ADDR, ecc_err);
1164	xsdfec_regwrite(xsdfec, XSDFEC_ISR_ADDR, isr_err);
1165
1166	tmp = ecc_err & XSDFEC_ALL_ECC_ISR_MBE_MASK;
1167	/* Count uncorrectable 2-bit errors */
1168	uecc_count = hweight32(tmp);
1169	/* Count all ECC errors */
1170	aecc_count = hweight32(ecc_err);
1171	/* Number of correctable 1-bit ECC error */
1172	cecc_count = aecc_count - 2 * uecc_count;
1173	/* Count ISR errors */
1174	isr_err_count = hweight32(isr_err);
1175	dev_dbg(xsdfec->dev, "tmp=%x, uecc=%x, aecc=%x, cecc=%x, isr=%x", tmp,
1176		uecc_count, aecc_count, cecc_count, isr_err_count);
1177	dev_dbg(xsdfec->dev, "uecc=%x, cecc=%x, isr=%x", xsdfec->uecc_count,
1178		xsdfec->cecc_count, xsdfec->isr_err_count);
1179
1180	spin_lock_irqsave(&xsdfec->error_data_lock, xsdfec->flags);
1181	/* Add new errors to a 2-bits counter */
1182	if (uecc_count)
1183		xsdfec->uecc_count += uecc_count;
1184	/* Add new errors to a 1-bits counter */
1185	if (cecc_count)
1186		xsdfec->cecc_count += cecc_count;
1187	/* Add new errors to a ISR counter */
1188	if (isr_err_count)
1189		xsdfec->isr_err_count += isr_err_count;
1190
1191	/* Update state/stats flag */
1192	if (uecc_count) {
1193		if (ecc_err & XSDFEC_ECC_ISR_MBE_MASK)
1194			xsdfec->state = XSDFEC_NEEDS_RESET;
1195		else if (ecc_err & XSDFEC_PL_INIT_ECC_ISR_MBE_MASK)
1196			xsdfec->state = XSDFEC_PL_RECONFIGURE;
1197		xsdfec->stats_updated = true;
1198		xsdfec->state_updated = true;
1199	}
1200
1201	if (cecc_count)
1202		xsdfec->stats_updated = true;
1203
1204	if (isr_err_count) {
1205		xsdfec->state = XSDFEC_NEEDS_RESET;
1206		xsdfec->stats_updated = true;
1207		xsdfec->state_updated = true;
1208	}
1209
1210	spin_unlock_irqrestore(&xsdfec->error_data_lock, xsdfec->flags);
1211	dev_dbg(xsdfec->dev, "state=%x, stats=%x", xsdfec->state_updated,
1212		xsdfec->stats_updated);
1213
1214	/* Enable another polling */
1215	if (xsdfec->state_updated || xsdfec->stats_updated)
1216		wake_up_interruptible(&xsdfec->waitq);
1217	else
1218		ret = IRQ_NONE;
1219
1220	/* Unmask Interrupts */
1221	xsdfec_isr_enable(xsdfec, true);
1222	xsdfec_ecc_isr_enable(xsdfec, true);
1223
1224	return ret;
1225}
1226
1227static int xsdfec_clk_init(struct platform_device *pdev,
1228			   struct xsdfec_clks *clks)
1229{
1230	int err;
1231
1232	clks->core_clk = devm_clk_get(&pdev->dev, "core_clk");
1233	if (IS_ERR(clks->core_clk)) {
1234		dev_err(&pdev->dev, "failed to get core_clk");
1235		return PTR_ERR(clks->core_clk);
1236	}
1237
1238	clks->axi_clk = devm_clk_get(&pdev->dev, "s_axi_aclk");
1239	if (IS_ERR(clks->axi_clk)) {
1240		dev_err(&pdev->dev, "failed to get axi_clk");
1241		return PTR_ERR(clks->axi_clk);
1242	}
1243
1244	clks->din_words_clk = devm_clk_get(&pdev->dev, "s_axis_din_words_aclk");
1245	if (IS_ERR(clks->din_words_clk)) {
1246		if (PTR_ERR(clks->din_words_clk) != -ENOENT) {
1247			err = PTR_ERR(clks->din_words_clk);
1248			return err;
1249		}
1250		clks->din_words_clk = NULL;
1251	}
1252
1253	clks->din_clk = devm_clk_get(&pdev->dev, "s_axis_din_aclk");
1254	if (IS_ERR(clks->din_clk)) {
1255		if (PTR_ERR(clks->din_clk) != -ENOENT) {
1256			err = PTR_ERR(clks->din_clk);
1257			return err;
1258		}
1259		clks->din_clk = NULL;
1260	}
1261
1262	clks->dout_clk = devm_clk_get(&pdev->dev, "m_axis_dout_aclk");
1263	if (IS_ERR(clks->dout_clk)) {
1264		if (PTR_ERR(clks->dout_clk) != -ENOENT) {
1265			err = PTR_ERR(clks->dout_clk);
1266			return err;
1267		}
1268		clks->dout_clk = NULL;
1269	}
1270
1271	clks->dout_words_clk =
1272		devm_clk_get(&pdev->dev, "s_axis_dout_words_aclk");
1273	if (IS_ERR(clks->dout_words_clk)) {
1274		if (PTR_ERR(clks->dout_words_clk) != -ENOENT) {
1275			err = PTR_ERR(clks->dout_words_clk);
1276			return err;
1277		}
1278		clks->dout_words_clk = NULL;
1279	}
1280
1281	clks->ctrl_clk = devm_clk_get(&pdev->dev, "s_axis_ctrl_aclk");
1282	if (IS_ERR(clks->ctrl_clk)) {
1283		if (PTR_ERR(clks->ctrl_clk) != -ENOENT) {
1284			err = PTR_ERR(clks->ctrl_clk);
1285			return err;
1286		}
1287		clks->ctrl_clk = NULL;
1288	}
1289
1290	clks->status_clk = devm_clk_get(&pdev->dev, "m_axis_status_aclk");
1291	if (IS_ERR(clks->status_clk)) {
1292		if (PTR_ERR(clks->status_clk) != -ENOENT) {
1293			err = PTR_ERR(clks->status_clk);
1294			return err;
1295		}
1296		clks->status_clk = NULL;
1297	}
1298
1299	err = clk_prepare_enable(clks->core_clk);
1300	if (err) {
1301		dev_err(&pdev->dev, "failed to enable core_clk (%d)", err);
1302		return err;
1303	}
1304
1305	err = clk_prepare_enable(clks->axi_clk);
1306	if (err) {
1307		dev_err(&pdev->dev, "failed to enable axi_clk (%d)", err);
1308		goto err_disable_core_clk;
1309	}
1310
1311	err = clk_prepare_enable(clks->din_clk);
1312	if (err) {
1313		dev_err(&pdev->dev, "failed to enable din_clk (%d)", err);
1314		goto err_disable_axi_clk;
1315	}
1316
1317	err = clk_prepare_enable(clks->din_words_clk);
1318	if (err) {
1319		dev_err(&pdev->dev, "failed to enable din_words_clk (%d)", err);
1320		goto err_disable_din_clk;
1321	}
1322
1323	err = clk_prepare_enable(clks->dout_clk);
1324	if (err) {
1325		dev_err(&pdev->dev, "failed to enable dout_clk (%d)", err);
1326		goto err_disable_din_words_clk;
1327	}
1328
1329	err = clk_prepare_enable(clks->dout_words_clk);
1330	if (err) {
1331		dev_err(&pdev->dev, "failed to enable dout_words_clk (%d)",
1332			err);
1333		goto err_disable_dout_clk;
1334	}
1335
1336	err = clk_prepare_enable(clks->ctrl_clk);
1337	if (err) {
1338		dev_err(&pdev->dev, "failed to enable ctrl_clk (%d)", err);
1339		goto err_disable_dout_words_clk;
1340	}
1341
1342	err = clk_prepare_enable(clks->status_clk);
1343	if (err) {
1344		dev_err(&pdev->dev, "failed to enable status_clk (%d)\n", err);
1345		goto err_disable_ctrl_clk;
1346	}
1347
1348	return err;
1349
1350err_disable_ctrl_clk:
1351	clk_disable_unprepare(clks->ctrl_clk);
1352err_disable_dout_words_clk:
1353	clk_disable_unprepare(clks->dout_words_clk);
1354err_disable_dout_clk:
1355	clk_disable_unprepare(clks->dout_clk);
1356err_disable_din_words_clk:
1357	clk_disable_unprepare(clks->din_words_clk);
1358err_disable_din_clk:
1359	clk_disable_unprepare(clks->din_clk);
1360err_disable_axi_clk:
1361	clk_disable_unprepare(clks->axi_clk);
1362err_disable_core_clk:
1363	clk_disable_unprepare(clks->core_clk);
1364
1365	return err;
1366}
1367
1368static void xsdfec_disable_all_clks(struct xsdfec_clks *clks)
1369{
1370	clk_disable_unprepare(clks->status_clk);
1371	clk_disable_unprepare(clks->ctrl_clk);
1372	clk_disable_unprepare(clks->dout_words_clk);
1373	clk_disable_unprepare(clks->dout_clk);
1374	clk_disable_unprepare(clks->din_words_clk);
1375	clk_disable_unprepare(clks->din_clk);
1376	clk_disable_unprepare(clks->core_clk);
1377	clk_disable_unprepare(clks->axi_clk);
1378}
1379
1380static int xsdfec_probe(struct platform_device *pdev)
1381{
1382	struct xsdfec_dev *xsdfec;
1383	struct device *dev;
1384	struct resource *res;
1385	int err;
1386	bool irq_enabled = true;
1387
1388	xsdfec = devm_kzalloc(&pdev->dev, sizeof(*xsdfec), GFP_KERNEL);
1389	if (!xsdfec)
1390		return -ENOMEM;
1391
1392	xsdfec->dev = &pdev->dev;
1393	spin_lock_init(&xsdfec->error_data_lock);
1394
1395	err = xsdfec_clk_init(pdev, &xsdfec->clks);
1396	if (err)
1397		return err;
1398
1399	dev = xsdfec->dev;
1400	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1401	xsdfec->regs = devm_ioremap_resource(dev, res);
1402	if (IS_ERR(xsdfec->regs)) {
1403		err = PTR_ERR(xsdfec->regs);
1404		goto err_xsdfec_dev;
1405	}
1406
1407	xsdfec->irq = platform_get_irq(pdev, 0);
1408	if (xsdfec->irq < 0) {
1409		dev_dbg(dev, "platform_get_irq failed");
1410		irq_enabled = false;
1411	}
1412
1413	err = xsdfec_parse_of(xsdfec);
1414	if (err < 0)
1415		goto err_xsdfec_dev;
1416
1417	update_config_from_hw(xsdfec);
1418
1419	/* Save driver private data */
1420	platform_set_drvdata(pdev, xsdfec);
1421
1422	if (irq_enabled) {
1423		init_waitqueue_head(&xsdfec->waitq);
1424		/* Register IRQ thread */
1425		err = devm_request_threaded_irq(dev, xsdfec->irq, NULL,
1426						xsdfec_irq_thread, IRQF_ONESHOT,
1427						"xilinx-sdfec16", xsdfec);
1428		if (err < 0) {
1429			dev_err(dev, "unable to request IRQ%d", xsdfec->irq);
1430			goto err_xsdfec_dev;
1431		}
1432	}
1433
1434	err = ida_alloc(&dev_nrs, GFP_KERNEL);
1435	if (err < 0)
1436		goto err_xsdfec_dev;
1437	xsdfec->dev_id = err;
1438
1439	snprintf(xsdfec->dev_name, DEV_NAME_LEN, "xsdfec%d", xsdfec->dev_id);
1440	xsdfec->miscdev.minor = MISC_DYNAMIC_MINOR;
1441	xsdfec->miscdev.name = xsdfec->dev_name;
1442	xsdfec->miscdev.fops = &xsdfec_fops;
1443	xsdfec->miscdev.parent = dev;
1444	err = misc_register(&xsdfec->miscdev);
1445	if (err) {
1446		dev_err(dev, "error:%d. Unable to register device", err);
1447		goto err_xsdfec_ida;
1448	}
1449	return 0;
1450
1451err_xsdfec_ida:
1452	ida_free(&dev_nrs, xsdfec->dev_id);
1453err_xsdfec_dev:
1454	xsdfec_disable_all_clks(&xsdfec->clks);
1455	return err;
1456}
1457
1458static int xsdfec_remove(struct platform_device *pdev)
1459{
1460	struct xsdfec_dev *xsdfec;
1461
1462	xsdfec = platform_get_drvdata(pdev);
1463	misc_deregister(&xsdfec->miscdev);
1464	ida_free(&dev_nrs, xsdfec->dev_id);
1465	xsdfec_disable_all_clks(&xsdfec->clks);
1466	return 0;
1467}
1468
1469static const struct of_device_id xsdfec_of_match[] = {
1470	{
1471		.compatible = "xlnx,sd-fec-1.1",
1472	},
1473	{ /* end of table */ }
1474};
1475MODULE_DEVICE_TABLE(of, xsdfec_of_match);
1476
1477static struct platform_driver xsdfec_driver = {
1478	.driver = {
1479		.name = "xilinx-sdfec",
1480		.of_match_table = xsdfec_of_match,
1481	},
1482	.probe = xsdfec_probe,
1483	.remove =  xsdfec_remove,
1484};
1485
1486module_platform_driver(xsdfec_driver);
1487
1488MODULE_AUTHOR("Xilinx, Inc");
1489MODULE_DESCRIPTION("Xilinx SD-FEC16 Driver");
1490MODULE_LICENSE("GPL");
1491