1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Device driver for the SYMBIOS/LSILOGIC 53C8XX and 53C1010 family
4 * of PCI-SCSI IO processors.
5 *
6 * Copyright (C) 1999-2001  Gerard Roudier <groudier@free.fr>
7 * Copyright (c) 2003-2005  Matthew Wilcox <matthew@wil.cx>
8 *
9 * This driver is derived from the Linux sym53c8xx driver.
10 * Copyright (C) 1998-2000  Gerard Roudier
11 *
12 * The sym53c8xx driver is derived from the ncr53c8xx driver that had been
13 * a port of the FreeBSD ncr driver to Linux-1.2.13.
14 *
15 * The original ncr driver has been written for 386bsd and FreeBSD by
16 *         Wolfgang Stanglmeier        <wolf@cologne.de>
17 *         Stefan Esser                <se@mi.Uni-Koeln.de>
18 * Copyright (C) 1994  Wolfgang Stanglmeier
19 *
20 * Other major contributions:
21 *
22 * NVRAM detection and reading.
23 * Copyright (C) 1997 Richard Waltham <dormouse@farsrobt.demon.co.uk>
24 *
25 *-----------------------------------------------------------------------------
26 */
27#include <linux/ctype.h>
28#include <linux/init.h>
29#include <linux/module.h>
30#include <linux/moduleparam.h>
31#include <linux/spinlock.h>
32#include <scsi/scsi.h>
33#include <scsi/scsi_tcq.h>
34#include <scsi/scsi_device.h>
35#include <scsi/scsi_transport.h>
36
37#include "sym_glue.h"
38#include "sym_nvram.h"
39
40#define NAME53C		"sym53c"
41#define NAME53C8XX	"sym53c8xx"
42
43struct sym_driver_setup sym_driver_setup = SYM_LINUX_DRIVER_SETUP;
44unsigned int sym_debug_flags = 0;
45
46static char *excl_string;
47static char *safe_string;
48module_param_named(cmd_per_lun, sym_driver_setup.max_tag, ushort, 0);
49module_param_named(burst, sym_driver_setup.burst_order, byte, 0);
50module_param_named(led, sym_driver_setup.scsi_led, byte, 0);
51module_param_named(diff, sym_driver_setup.scsi_diff, byte, 0);
52module_param_named(irqm, sym_driver_setup.irq_mode, byte, 0);
53module_param_named(buschk, sym_driver_setup.scsi_bus_check, byte, 0);
54module_param_named(hostid, sym_driver_setup.host_id, byte, 0);
55module_param_named(verb, sym_driver_setup.verbose, byte, 0);
56module_param_named(debug, sym_debug_flags, uint, 0);
57module_param_named(settle, sym_driver_setup.settle_delay, byte, 0);
58module_param_named(nvram, sym_driver_setup.use_nvram, byte, 0);
59module_param_named(excl, excl_string, charp, 0);
60module_param_named(safe, safe_string, charp, 0);
61
62MODULE_PARM_DESC(cmd_per_lun, "The maximum number of tags to use by default");
63MODULE_PARM_DESC(burst, "Maximum burst.  0 to disable, 255 to read from registers");
64MODULE_PARM_DESC(led, "Set to 1 to enable LED support");
65MODULE_PARM_DESC(diff, "0 for no differential mode, 1 for BIOS, 2 for always, 3 for not GPIO3");
66MODULE_PARM_DESC(irqm, "0 for open drain, 1 to leave alone, 2 for totem pole");
67MODULE_PARM_DESC(buschk, "0 to not check, 1 for detach on error, 2 for warn on error");
68MODULE_PARM_DESC(hostid, "The SCSI ID to use for the host adapters");
69MODULE_PARM_DESC(verb, "0 for minimal verbosity, 1 for normal, 2 for excessive");
70MODULE_PARM_DESC(debug, "Set bits to enable debugging");
71MODULE_PARM_DESC(settle, "Settle delay in seconds.  Default 3");
72MODULE_PARM_DESC(nvram, "Option currently not used");
73MODULE_PARM_DESC(excl, "List ioport addresses here to prevent controllers from being attached");
74MODULE_PARM_DESC(safe, "Set other settings to a \"safe mode\"");
75
76MODULE_LICENSE("GPL");
77MODULE_VERSION(SYM_VERSION);
78MODULE_AUTHOR("Matthew Wilcox <matthew@wil.cx>");
79MODULE_DESCRIPTION("NCR, Symbios and LSI 8xx and 1010 PCI SCSI adapters");
80
81static void sym2_setup_params(void)
82{
83	char *p = excl_string;
84	int xi = 0;
85
86	while (p && (xi < 8)) {
87		char *next_p;
88		int val = (int) simple_strtoul(p, &next_p, 0);
89		sym_driver_setup.excludes[xi++] = val;
90		p = next_p;
91	}
92
93	if (safe_string) {
94		if (*safe_string == 'y') {
95			sym_driver_setup.max_tag = 0;
96			sym_driver_setup.burst_order = 0;
97			sym_driver_setup.scsi_led = 0;
98			sym_driver_setup.scsi_diff = 1;
99			sym_driver_setup.irq_mode = 0;
100			sym_driver_setup.scsi_bus_check = 2;
101			sym_driver_setup.host_id = 7;
102			sym_driver_setup.verbose = 2;
103			sym_driver_setup.settle_delay = 10;
104			sym_driver_setup.use_nvram = 1;
105		} else if (*safe_string != 'n') {
106			printk(KERN_WARNING NAME53C8XX "Ignoring parameter %s"
107					" passed to safe option", safe_string);
108		}
109	}
110}
111
112static struct scsi_transport_template *sym2_transport_template = NULL;
113
114/*
115 *  Driver private area in the SCSI command structure.
116 */
117struct sym_ucmd {		/* Override the SCSI pointer structure */
118	struct completion *eh_done;		/* SCSI error handling */
119};
120
121#define SYM_UCMD_PTR(cmd)  ((struct sym_ucmd *)scsi_cmd_priv(cmd))
122#define SYM_SOFTC_PTR(cmd) sym_get_hcb(cmd->device->host)
123
124/*
125 *  Complete a pending CAM CCB.
126 */
127void sym_xpt_done(struct sym_hcb *np, struct scsi_cmnd *cmd)
128{
129	struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd);
130
131	if (ucmd->eh_done)
132		complete(ucmd->eh_done);
133
134	scsi_dma_unmap(cmd);
135	scsi_done(cmd);
136}
137
138/*
139 *  Tell the SCSI layer about a BUS RESET.
140 */
141void sym_xpt_async_bus_reset(struct sym_hcb *np)
142{
143	printf_notice("%s: SCSI BUS has been reset.\n", sym_name(np));
144	np->s.settle_time = jiffies + sym_driver_setup.settle_delay * HZ;
145	np->s.settle_time_valid = 1;
146	if (sym_verbose >= 2)
147		printf_info("%s: command processing suspended for %d seconds\n",
148			    sym_name(np), sym_driver_setup.settle_delay);
149}
150
151/*
152 *  Choose the more appropriate CAM status if
153 *  the IO encountered an extended error.
154 */
155static int sym_xerr_cam_status(int cam_status, int x_status)
156{
157	if (x_status) {
158		if (x_status & XE_PARITY_ERR)
159			cam_status = DID_PARITY;
160		else
161			cam_status = DID_ERROR;
162	}
163	return cam_status;
164}
165
166/*
167 *  Build CAM result for a failed or auto-sensed IO.
168 */
169void sym_set_cam_result_error(struct sym_hcb *np, struct sym_ccb *cp, int resid)
170{
171	struct scsi_cmnd *cmd = cp->cmd;
172	u_int cam_status, scsi_status;
173
174	cam_status  = DID_OK;
175	scsi_status = cp->ssss_status;
176
177	if (cp->host_flags & HF_SENSE) {
178		scsi_status = cp->sv_scsi_status;
179		resid = cp->sv_resid;
180		if (sym_verbose && cp->sv_xerr_status)
181			sym_print_xerr(cmd, cp->sv_xerr_status);
182		if (cp->host_status == HS_COMPLETE &&
183		    cp->ssss_status == S_GOOD &&
184		    cp->xerr_status == 0) {
185			cam_status = sym_xerr_cam_status(DID_OK,
186							 cp->sv_xerr_status);
187			/*
188			 *  Bounce back the sense data to user.
189			 */
190			memset(cmd->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE);
191			memcpy(cmd->sense_buffer, cp->sns_bbuf,
192			       min(SCSI_SENSE_BUFFERSIZE, SYM_SNS_BBUF_LEN));
193#if 0
194			/*
195			 *  If the device reports a UNIT ATTENTION condition
196			 *  due to a RESET condition, we should consider all
197			 *  disconnect CCBs for this unit as aborted.
198			 */
199			if (1) {
200				u_char *p;
201				p  = (u_char *) cmd->sense_data;
202				if (p[0]==0x70 && p[2]==0x6 && p[12]==0x29)
203					sym_clear_tasks(np, DID_ABORT,
204							cp->target,cp->lun, -1);
205			}
206#endif
207		} else {
208			/*
209			 * Error return from our internal request sense.  This
210			 * is bad: we must clear the contingent allegiance
211			 * condition otherwise the device will always return
212			 * BUSY.  Use a big stick.
213			 */
214			sym_reset_scsi_target(np, cmd->device->id);
215			cam_status = DID_ERROR;
216		}
217	} else if (cp->host_status == HS_COMPLETE) 	/* Bad SCSI status */
218		cam_status = DID_OK;
219	else if (cp->host_status == HS_SEL_TIMEOUT)	/* Selection timeout */
220		cam_status = DID_NO_CONNECT;
221	else if (cp->host_status == HS_UNEXPECTED)	/* Unexpected BUS FREE*/
222		cam_status = DID_ERROR;
223	else {						/* Extended error */
224		if (sym_verbose) {
225			sym_print_addr(cmd, "COMMAND FAILED (%x %x %x).\n",
226				cp->host_status, cp->ssss_status,
227				cp->xerr_status);
228		}
229		/*
230		 *  Set the most appropriate value for CAM status.
231		 */
232		cam_status = sym_xerr_cam_status(DID_ERROR, cp->xerr_status);
233	}
234	scsi_set_resid(cmd, resid);
235	cmd->result = (cam_status << 16) | scsi_status;
236}
237
238static int sym_scatter(struct sym_hcb *np, struct sym_ccb *cp, struct scsi_cmnd *cmd)
239{
240	int segment;
241	int use_sg;
242
243	cp->data_len = 0;
244
245	use_sg = scsi_dma_map(cmd);
246	if (use_sg > 0) {
247		struct scatterlist *sg;
248		struct sym_tcb *tp = &np->target[cp->target];
249		struct sym_tblmove *data;
250
251		if (use_sg > SYM_CONF_MAX_SG) {
252			scsi_dma_unmap(cmd);
253			return -1;
254		}
255
256		data = &cp->phys.data[SYM_CONF_MAX_SG - use_sg];
257
258		scsi_for_each_sg(cmd, sg, use_sg, segment) {
259			dma_addr_t baddr = sg_dma_address(sg);
260			unsigned int len = sg_dma_len(sg);
261
262			if ((len & 1) && (tp->head.wval & EWS)) {
263				len++;
264				cp->odd_byte_adjustment++;
265			}
266
267			sym_build_sge(np, &data[segment], baddr, len);
268			cp->data_len += len;
269		}
270	} else {
271		segment = -2;
272	}
273
274	return segment;
275}
276
277/*
278 *  Queue a SCSI command.
279 */
280static int sym_queue_command(struct sym_hcb *np, struct scsi_cmnd *cmd)
281{
282	struct scsi_device *sdev = cmd->device;
283	struct sym_tcb *tp;
284	struct sym_lcb *lp;
285	struct sym_ccb *cp;
286	int	order;
287
288	/*
289	 *  Retrieve the target descriptor.
290	 */
291	tp = &np->target[sdev->id];
292
293	/*
294	 *  Select tagged/untagged.
295	 */
296	lp = sym_lp(tp, sdev->lun);
297	order = (lp && lp->s.reqtags) ? M_SIMPLE_TAG : 0;
298
299	/*
300	 *  Queue the SCSI IO.
301	 */
302	cp = sym_get_ccb(np, cmd, order);
303	if (!cp)
304		return 1;	/* Means resource shortage */
305	sym_queue_scsiio(np, cmd, cp);
306	return 0;
307}
308
309/*
310 *  Setup buffers and pointers that address the CDB.
311 */
312static inline int sym_setup_cdb(struct sym_hcb *np, struct scsi_cmnd *cmd, struct sym_ccb *cp)
313{
314	memcpy(cp->cdb_buf, cmd->cmnd, cmd->cmd_len);
315
316	cp->phys.cmd.addr = CCB_BA(cp, cdb_buf[0]);
317	cp->phys.cmd.size = cpu_to_scr(cmd->cmd_len);
318
319	return 0;
320}
321
322/*
323 *  Setup pointers that address the data and start the I/O.
324 */
325int sym_setup_data_and_start(struct sym_hcb *np, struct scsi_cmnd *cmd, struct sym_ccb *cp)
326{
327	u32 lastp, goalp;
328	int dir;
329
330	/*
331	 *  Build the CDB.
332	 */
333	if (sym_setup_cdb(np, cmd, cp))
334		goto out_abort;
335
336	/*
337	 *  No direction means no data.
338	 */
339	dir = cmd->sc_data_direction;
340	if (dir != DMA_NONE) {
341		cp->segments = sym_scatter(np, cp, cmd);
342		if (cp->segments < 0) {
343			sym_set_cam_status(cmd, DID_ERROR);
344			goto out_abort;
345		}
346
347		/*
348		 *  No segments means no data.
349		 */
350		if (!cp->segments)
351			dir = DMA_NONE;
352	} else {
353		cp->data_len = 0;
354		cp->segments = 0;
355	}
356
357	/*
358	 *  Set the data pointer.
359	 */
360	switch (dir) {
361	case DMA_BIDIRECTIONAL:
362		scmd_printk(KERN_INFO, cmd, "got DMA_BIDIRECTIONAL command");
363		sym_set_cam_status(cmd, DID_ERROR);
364		goto out_abort;
365	case DMA_TO_DEVICE:
366		goalp = SCRIPTA_BA(np, data_out2) + 8;
367		lastp = goalp - 8 - (cp->segments * (2*4));
368		break;
369	case DMA_FROM_DEVICE:
370		cp->host_flags |= HF_DATA_IN;
371		goalp = SCRIPTA_BA(np, data_in2) + 8;
372		lastp = goalp - 8 - (cp->segments * (2*4));
373		break;
374	case DMA_NONE:
375	default:
376		lastp = goalp = SCRIPTB_BA(np, no_data);
377		break;
378	}
379
380	/*
381	 *  Set all pointers values needed by SCRIPTS.
382	 */
383	cp->phys.head.lastp = cpu_to_scr(lastp);
384	cp->phys.head.savep = cpu_to_scr(lastp);
385	cp->startp	    = cp->phys.head.savep;
386	cp->goalp	    = cpu_to_scr(goalp);
387
388	/*
389	 *  When `#ifed 1', the code below makes the driver
390	 *  panic on the first attempt to write to a SCSI device.
391	 *  It is the first test we want to do after a driver
392	 *  change that does not seem obviously safe. :)
393	 */
394#if 0
395	switch (cp->cdb_buf[0]) {
396	case 0x0A: case 0x2A: case 0xAA:
397		panic("XXXXXXXXXXXXX WRITE NOT YET ALLOWED XXXXXXXXXXXXXX\n");
398		break;
399	default:
400		break;
401	}
402#endif
403
404	/*
405	 *	activate this job.
406	 */
407	sym_put_start_queue(np, cp);
408	return 0;
409
410out_abort:
411	sym_free_ccb(np, cp);
412	sym_xpt_done(np, cmd);
413	return 0;
414}
415
416
417/*
418 *  timer daemon.
419 *
420 *  Misused to keep the driver running when
421 *  interrupts are not configured correctly.
422 */
423static void sym_timer(struct sym_hcb *np)
424{
425	unsigned long thistime = jiffies;
426
427	/*
428	 *  Restart the timer.
429	 */
430	np->s.timer.expires = thistime + SYM_CONF_TIMER_INTERVAL;
431	add_timer(&np->s.timer);
432
433	/*
434	 *  If we are resetting the ncr, wait for settle_time before
435	 *  clearing it. Then command processing will be resumed.
436	 */
437	if (np->s.settle_time_valid) {
438		if (time_before_eq(np->s.settle_time, thistime)) {
439			if (sym_verbose >= 2 )
440				printk("%s: command processing resumed\n",
441				       sym_name(np));
442			np->s.settle_time_valid = 0;
443		}
444		return;
445	}
446
447	/*
448	 *	Nothing to do for now, but that may come.
449	 */
450	if (np->s.lasttime + 4*HZ < thistime) {
451		np->s.lasttime = thistime;
452	}
453
454#ifdef SYM_CONF_PCIQ_MAY_MISS_COMPLETIONS
455	/*
456	 *  Some way-broken PCI bridges may lead to
457	 *  completions being lost when the clearing
458	 *  of the INTFLY flag by the CPU occurs
459	 *  concurrently with the chip raising this flag.
460	 *  If this ever happen, lost completions will
461	 * be reaped here.
462	 */
463	sym_wakeup_done(np);
464#endif
465}
466
467
468/*
469 *  PCI BUS error handler.
470 */
471void sym_log_bus_error(struct Scsi_Host *shost)
472{
473	struct sym_data *sym_data = shost_priv(shost);
474	struct pci_dev *pdev = sym_data->pdev;
475	unsigned short pci_sts;
476	pci_read_config_word(pdev, PCI_STATUS, &pci_sts);
477	if (pci_sts & 0xf900) {
478		pci_write_config_word(pdev, PCI_STATUS, pci_sts);
479		shost_printk(KERN_WARNING, shost,
480			"PCI bus error: status = 0x%04x\n", pci_sts & 0xf900);
481	}
482}
483
484/*
485 * queuecommand method.  Entered with the host adapter lock held and
486 * interrupts disabled.
487 */
488static int sym53c8xx_queue_command_lck(struct scsi_cmnd *cmd)
489{
490	struct sym_hcb *np = SYM_SOFTC_PTR(cmd);
491	struct sym_ucmd *ucp = SYM_UCMD_PTR(cmd);
492	int sts = 0;
493
494	memset(ucp, 0, sizeof(*ucp));
495
496	/*
497	 *  Shorten our settle_time if needed for
498	 *  this command not to time out.
499	 */
500	if (np->s.settle_time_valid && scsi_cmd_to_rq(cmd)->timeout) {
501		unsigned long tlimit = jiffies + scsi_cmd_to_rq(cmd)->timeout;
502		tlimit -= SYM_CONF_TIMER_INTERVAL*2;
503		if (time_after(np->s.settle_time, tlimit)) {
504			np->s.settle_time = tlimit;
505		}
506	}
507
508	if (np->s.settle_time_valid)
509		return SCSI_MLQUEUE_HOST_BUSY;
510
511	sts = sym_queue_command(np, cmd);
512	if (sts)
513		return SCSI_MLQUEUE_HOST_BUSY;
514	return 0;
515}
516
517static DEF_SCSI_QCMD(sym53c8xx_queue_command)
518
519/*
520 *  Linux entry point of the interrupt handler.
521 */
522static irqreturn_t sym53c8xx_intr(int irq, void *dev_id)
523{
524	struct Scsi_Host *shost = dev_id;
525	struct sym_data *sym_data = shost_priv(shost);
526	irqreturn_t result;
527
528	/* Avoid spinloop trying to handle interrupts on frozen device */
529	if (pci_channel_offline(sym_data->pdev))
530		return IRQ_NONE;
531
532	if (DEBUG_FLAGS & DEBUG_TINY) printf_debug ("[");
533
534	spin_lock(shost->host_lock);
535	result = sym_interrupt(shost);
536	spin_unlock(shost->host_lock);
537
538	if (DEBUG_FLAGS & DEBUG_TINY) printf_debug ("]\n");
539
540	return result;
541}
542
543/*
544 *  Linux entry point of the timer handler
545 */
546static void sym53c8xx_timer(struct timer_list *t)
547{
548	struct sym_hcb *np = from_timer(np, t, s.timer);
549	unsigned long flags;
550
551	spin_lock_irqsave(np->s.host->host_lock, flags);
552	sym_timer(np);
553	spin_unlock_irqrestore(np->s.host->host_lock, flags);
554}
555
556
557/*
558 *  What the eh thread wants us to perform.
559 */
560#define SYM_EH_ABORT		0
561#define SYM_EH_DEVICE_RESET	1
562#define SYM_EH_BUS_RESET	2
563#define SYM_EH_HOST_RESET	3
564
565/*
566 *  Generic method for our eh processing.
567 *  The 'op' argument tells what we have to do.
568 */
569static int sym_eh_handler(int op, char *opname, struct scsi_cmnd *cmd)
570{
571	struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd);
572	struct Scsi_Host *shost = cmd->device->host;
573	struct sym_data *sym_data = shost_priv(shost);
574	struct pci_dev *pdev = sym_data->pdev;
575	struct sym_hcb *np = sym_data->ncb;
576	SYM_QUEHEAD *qp;
577	int cmd_queued = 0;
578	int sts = -1;
579	struct completion eh_done;
580
581	scmd_printk(KERN_WARNING, cmd, "%s operation started\n", opname);
582
583	/* We may be in an error condition because the PCI bus
584	 * went down. In this case, we need to wait until the
585	 * PCI bus is reset, the card is reset, and only then
586	 * proceed with the scsi error recovery.  There's no
587	 * point in hurrying; take a leisurely wait.
588	 */
589#define WAIT_FOR_PCI_RECOVERY	35
590	if (pci_channel_offline(pdev)) {
591		int finished_reset = 0;
592		init_completion(&eh_done);
593		spin_lock_irq(shost->host_lock);
594		/* Make sure we didn't race */
595		if (pci_channel_offline(pdev)) {
596			BUG_ON(sym_data->io_reset);
597			sym_data->io_reset = &eh_done;
598		} else {
599			finished_reset = 1;
600		}
601		spin_unlock_irq(shost->host_lock);
602		if (!finished_reset)
603			finished_reset = wait_for_completion_timeout
604						(sym_data->io_reset,
605						WAIT_FOR_PCI_RECOVERY*HZ);
606		spin_lock_irq(shost->host_lock);
607		sym_data->io_reset = NULL;
608		spin_unlock_irq(shost->host_lock);
609		if (!finished_reset)
610			return SCSI_FAILED;
611	}
612
613	spin_lock_irq(shost->host_lock);
614	/* This one is queued in some place -> to wait for completion */
615	FOR_EACH_QUEUED_ELEMENT(&np->busy_ccbq, qp) {
616		struct sym_ccb *cp = sym_que_entry(qp, struct sym_ccb, link_ccbq);
617		if (cp->cmd == cmd) {
618			cmd_queued = 1;
619			break;
620		}
621	}
622
623	/* Try to proceed the operation we have been asked for */
624	sts = -1;
625	switch(op) {
626	case SYM_EH_ABORT:
627		sts = sym_abort_scsiio(np, cmd, 1);
628		break;
629	case SYM_EH_DEVICE_RESET:
630		sts = sym_reset_scsi_target(np, cmd->device->id);
631		break;
632	case SYM_EH_BUS_RESET:
633		sym_reset_scsi_bus(np, 1);
634		sts = 0;
635		break;
636	case SYM_EH_HOST_RESET:
637		sym_reset_scsi_bus(np, 0);
638		sym_start_up(shost, 1);
639		sts = 0;
640		break;
641	default:
642		break;
643	}
644
645	/* On error, restore everything and cross fingers :) */
646	if (sts)
647		cmd_queued = 0;
648
649	if (cmd_queued) {
650		init_completion(&eh_done);
651		ucmd->eh_done = &eh_done;
652		spin_unlock_irq(shost->host_lock);
653		if (!wait_for_completion_timeout(&eh_done, 5*HZ)) {
654			ucmd->eh_done = NULL;
655			sts = -2;
656		}
657	} else {
658		spin_unlock_irq(shost->host_lock);
659	}
660
661	dev_warn(&cmd->device->sdev_gendev, "%s operation %s.\n", opname,
662			sts==0 ? "complete" :sts==-2 ? "timed-out" : "failed");
663	return sts ? SCSI_FAILED : SCSI_SUCCESS;
664}
665
666
667/*
668 * Error handlers called from the eh thread (one thread per HBA).
669 */
670static int sym53c8xx_eh_abort_handler(struct scsi_cmnd *cmd)
671{
672	return sym_eh_handler(SYM_EH_ABORT, "ABORT", cmd);
673}
674
675static int sym53c8xx_eh_device_reset_handler(struct scsi_cmnd *cmd)
676{
677	return sym_eh_handler(SYM_EH_DEVICE_RESET, "DEVICE RESET", cmd);
678}
679
680static int sym53c8xx_eh_bus_reset_handler(struct scsi_cmnd *cmd)
681{
682	return sym_eh_handler(SYM_EH_BUS_RESET, "BUS RESET", cmd);
683}
684
685static int sym53c8xx_eh_host_reset_handler(struct scsi_cmnd *cmd)
686{
687	return sym_eh_handler(SYM_EH_HOST_RESET, "HOST RESET", cmd);
688}
689
690/*
691 *  Tune device queuing depth, according to various limits.
692 */
693static void sym_tune_dev_queuing(struct sym_tcb *tp, int lun, u_short reqtags)
694{
695	struct sym_lcb *lp = sym_lp(tp, lun);
696	u_short	oldtags;
697
698	if (!lp)
699		return;
700
701	oldtags = lp->s.reqtags;
702
703	if (reqtags > lp->s.scdev_depth)
704		reqtags = lp->s.scdev_depth;
705
706	lp->s.reqtags     = reqtags;
707
708	if (reqtags != oldtags) {
709		dev_info(&tp->starget->dev,
710		         "tagged command queuing %s, command queue depth %d.\n",
711		          lp->s.reqtags ? "enabled" : "disabled", reqtags);
712	}
713}
714
715static int sym53c8xx_slave_alloc(struct scsi_device *sdev)
716{
717	struct sym_hcb *np = sym_get_hcb(sdev->host);
718	struct sym_tcb *tp = &np->target[sdev->id];
719	struct sym_lcb *lp;
720	unsigned long flags;
721	int error;
722
723	if (sdev->id >= SYM_CONF_MAX_TARGET || sdev->lun >= SYM_CONF_MAX_LUN)
724		return -ENXIO;
725
726	spin_lock_irqsave(np->s.host->host_lock, flags);
727
728	/*
729	 * Fail the device init if the device is flagged NOSCAN at BOOT in
730	 * the NVRAM.  This may speed up boot and maintain coherency with
731	 * BIOS device numbering.  Clearing the flag allows the user to
732	 * rescan skipped devices later.  We also return an error for
733	 * devices not flagged for SCAN LUNS in the NVRAM since some single
734	 * lun devices behave badly when asked for a non zero LUN.
735	 */
736
737	if (tp->usrflags & SYM_SCAN_BOOT_DISABLED) {
738		tp->usrflags &= ~SYM_SCAN_BOOT_DISABLED;
739		starget_printk(KERN_INFO, sdev->sdev_target,
740				"Scan at boot disabled in NVRAM\n");
741		error = -ENXIO;
742		goto out;
743	}
744
745	if (tp->usrflags & SYM_SCAN_LUNS_DISABLED) {
746		if (sdev->lun != 0) {
747			error = -ENXIO;
748			goto out;
749		}
750		starget_printk(KERN_INFO, sdev->sdev_target,
751				"Multiple LUNs disabled in NVRAM\n");
752	}
753
754	lp = sym_alloc_lcb(np, sdev->id, sdev->lun);
755	if (!lp) {
756		error = -ENOMEM;
757		goto out;
758	}
759	if (tp->nlcb == 1)
760		tp->starget = sdev->sdev_target;
761
762	spi_min_period(tp->starget) = tp->usr_period;
763	spi_max_width(tp->starget) = tp->usr_width;
764
765	error = 0;
766out:
767	spin_unlock_irqrestore(np->s.host->host_lock, flags);
768
769	return error;
770}
771
772/*
773 * Linux entry point for device queue sizing.
774 */
775static int sym53c8xx_slave_configure(struct scsi_device *sdev)
776{
777	struct sym_hcb *np = sym_get_hcb(sdev->host);
778	struct sym_tcb *tp = &np->target[sdev->id];
779	struct sym_lcb *lp = sym_lp(tp, sdev->lun);
780	int reqtags, depth_to_use;
781
782	/*
783	 *  Get user flags.
784	 */
785	lp->curr_flags = lp->user_flags;
786
787	/*
788	 *  Select queue depth from driver setup.
789	 *  Do not use more than configured by user.
790	 *  Use at least 1.
791	 *  Do not use more than our maximum.
792	 */
793	reqtags = sym_driver_setup.max_tag;
794	if (reqtags > tp->usrtags)
795		reqtags = tp->usrtags;
796	if (!sdev->tagged_supported)
797		reqtags = 0;
798	if (reqtags > SYM_CONF_MAX_TAG)
799		reqtags = SYM_CONF_MAX_TAG;
800	depth_to_use = reqtags ? reqtags : 1;
801	scsi_change_queue_depth(sdev, depth_to_use);
802	lp->s.scdev_depth = depth_to_use;
803	sym_tune_dev_queuing(tp, sdev->lun, reqtags);
804
805	if (!spi_initial_dv(sdev->sdev_target))
806		spi_dv_device(sdev);
807
808	return 0;
809}
810
811static void sym53c8xx_slave_destroy(struct scsi_device *sdev)
812{
813	struct sym_hcb *np = sym_get_hcb(sdev->host);
814	struct sym_tcb *tp = &np->target[sdev->id];
815	struct sym_lcb *lp = sym_lp(tp, sdev->lun);
816	unsigned long flags;
817
818	/* if slave_alloc returned before allocating a sym_lcb, return */
819	if (!lp)
820		return;
821
822	spin_lock_irqsave(np->s.host->host_lock, flags);
823
824	if (lp->busy_itlq || lp->busy_itl) {
825		/*
826		 * This really shouldn't happen, but we can't return an error
827		 * so let's try to stop all on-going I/O.
828		 */
829		starget_printk(KERN_WARNING, tp->starget,
830			       "Removing busy LCB (%d)\n", (u8)sdev->lun);
831		sym_reset_scsi_bus(np, 1);
832	}
833
834	if (sym_free_lcb(np, sdev->id, sdev->lun) == 0) {
835		/*
836		 * It was the last unit for this target.
837		 */
838		tp->head.sval        = 0;
839		tp->head.wval        = np->rv_scntl3;
840		tp->head.uval        = 0;
841		tp->tgoal.check_nego = 1;
842		tp->starget	     = NULL;
843	}
844
845	spin_unlock_irqrestore(np->s.host->host_lock, flags);
846}
847
848/*
849 *  Linux entry point for info() function
850 */
851static const char *sym53c8xx_info (struct Scsi_Host *host)
852{
853	return SYM_DRIVER_NAME;
854}
855
856
857#ifdef SYM_LINUX_PROC_INFO_SUPPORT
858/*
859 *  Proc file system stuff
860 *
861 *  A read operation returns adapter information.
862 *  A write operation is a control command.
863 *  The string is parsed in the driver code and the command is passed
864 *  to the sym_usercmd() function.
865 */
866
867#ifdef SYM_LINUX_USER_COMMAND_SUPPORT
868
869struct	sym_usrcmd {
870	u_long	target;
871	u_long	lun;
872	u_long	data;
873	u_long	cmd;
874};
875
876#define UC_SETSYNC      10
877#define UC_SETTAGS	11
878#define UC_SETDEBUG	12
879#define UC_SETWIDE	14
880#define UC_SETFLAG	15
881#define UC_SETVERBOSE	17
882#define UC_RESETDEV	18
883#define UC_CLEARDEV	19
884
885static void sym_exec_user_command (struct sym_hcb *np, struct sym_usrcmd *uc)
886{
887	struct sym_tcb *tp;
888	int t, l;
889
890	switch (uc->cmd) {
891	case 0: return;
892
893#ifdef SYM_LINUX_DEBUG_CONTROL_SUPPORT
894	case UC_SETDEBUG:
895		sym_debug_flags = uc->data;
896		break;
897#endif
898	case UC_SETVERBOSE:
899		np->verbose = uc->data;
900		break;
901	default:
902		/*
903		 * We assume that other commands apply to targets.
904		 * This should always be the case and avoid the below
905		 * 4 lines to be repeated 6 times.
906		 */
907		for (t = 0; t < SYM_CONF_MAX_TARGET; t++) {
908			if (!((uc->target >> t) & 1))
909				continue;
910			tp = &np->target[t];
911			if (!tp->nlcb)
912				continue;
913
914			switch (uc->cmd) {
915
916			case UC_SETSYNC:
917				if (!uc->data || uc->data >= 255) {
918					tp->tgoal.iu = tp->tgoal.dt =
919						tp->tgoal.qas = 0;
920					tp->tgoal.offset = 0;
921				} else if (uc->data <= 9 && np->minsync_dt) {
922					if (uc->data < np->minsync_dt)
923						uc->data = np->minsync_dt;
924					tp->tgoal.iu = tp->tgoal.dt =
925						tp->tgoal.qas = 1;
926					tp->tgoal.width = 1;
927					tp->tgoal.period = uc->data;
928					tp->tgoal.offset = np->maxoffs_dt;
929				} else {
930					if (uc->data < np->minsync)
931						uc->data = np->minsync;
932					tp->tgoal.iu = tp->tgoal.dt =
933						tp->tgoal.qas = 0;
934					tp->tgoal.period = uc->data;
935					tp->tgoal.offset = np->maxoffs;
936				}
937				tp->tgoal.check_nego = 1;
938				break;
939			case UC_SETWIDE:
940				tp->tgoal.width = uc->data ? 1 : 0;
941				tp->tgoal.check_nego = 1;
942				break;
943			case UC_SETTAGS:
944				for (l = 0; l < SYM_CONF_MAX_LUN; l++)
945					sym_tune_dev_queuing(tp, l, uc->data);
946				break;
947			case UC_RESETDEV:
948				tp->to_reset = 1;
949				np->istat_sem = SEM;
950				OUTB(np, nc_istat, SIGP|SEM);
951				break;
952			case UC_CLEARDEV:
953				for (l = 0; l < SYM_CONF_MAX_LUN; l++) {
954					struct sym_lcb *lp = sym_lp(tp, l);
955					if (lp) lp->to_clear = 1;
956				}
957				np->istat_sem = SEM;
958				OUTB(np, nc_istat, SIGP|SEM);
959				break;
960			case UC_SETFLAG:
961				tp->usrflags = uc->data;
962				break;
963			}
964		}
965		break;
966	}
967}
968
969static int sym_skip_spaces(char *ptr, int len)
970{
971	int cnt, c;
972
973	for (cnt = len; cnt > 0 && (c = *ptr++) && isspace(c); cnt--);
974
975	return (len - cnt);
976}
977
978static int get_int_arg(char *ptr, int len, u_long *pv)
979{
980	char *end;
981
982	*pv = simple_strtoul(ptr, &end, 10);
983	return (end - ptr);
984}
985
986static int is_keyword(char *ptr, int len, char *verb)
987{
988	int verb_len = strlen(verb);
989
990	if (len >= verb_len && !memcmp(verb, ptr, verb_len))
991		return verb_len;
992	else
993		return 0;
994}
995
996#define SKIP_SPACES(ptr, len)						\
997	if ((arg_len = sym_skip_spaces(ptr, len)) < 1)			\
998		return -EINVAL;						\
999	ptr += arg_len; len -= arg_len;
1000
1001#define GET_INT_ARG(ptr, len, v)					\
1002	if (!(arg_len = get_int_arg(ptr, len, &(v))))			\
1003		return -EINVAL;						\
1004	ptr += arg_len; len -= arg_len;
1005
1006
1007/*
1008 * Parse a control command
1009 */
1010
1011static int sym_user_command(struct Scsi_Host *shost, char *buffer, int length)
1012{
1013	struct sym_hcb *np = sym_get_hcb(shost);
1014	char *ptr	= buffer;
1015	int len		= length;
1016	struct sym_usrcmd cmd, *uc = &cmd;
1017	int		arg_len;
1018	u_long 		target;
1019
1020	memset(uc, 0, sizeof(*uc));
1021
1022	if (len > 0 && ptr[len-1] == '\n')
1023		--len;
1024
1025	if	((arg_len = is_keyword(ptr, len, "setsync")) != 0)
1026		uc->cmd = UC_SETSYNC;
1027	else if	((arg_len = is_keyword(ptr, len, "settags")) != 0)
1028		uc->cmd = UC_SETTAGS;
1029	else if	((arg_len = is_keyword(ptr, len, "setverbose")) != 0)
1030		uc->cmd = UC_SETVERBOSE;
1031	else if	((arg_len = is_keyword(ptr, len, "setwide")) != 0)
1032		uc->cmd = UC_SETWIDE;
1033#ifdef SYM_LINUX_DEBUG_CONTROL_SUPPORT
1034	else if	((arg_len = is_keyword(ptr, len, "setdebug")) != 0)
1035		uc->cmd = UC_SETDEBUG;
1036#endif
1037	else if	((arg_len = is_keyword(ptr, len, "setflag")) != 0)
1038		uc->cmd = UC_SETFLAG;
1039	else if	((arg_len = is_keyword(ptr, len, "resetdev")) != 0)
1040		uc->cmd = UC_RESETDEV;
1041	else if	((arg_len = is_keyword(ptr, len, "cleardev")) != 0)
1042		uc->cmd = UC_CLEARDEV;
1043	else
1044		arg_len = 0;
1045
1046#ifdef DEBUG_PROC_INFO
1047printk("sym_user_command: arg_len=%d, cmd=%ld\n", arg_len, uc->cmd);
1048#endif
1049
1050	if (!arg_len)
1051		return -EINVAL;
1052	ptr += arg_len; len -= arg_len;
1053
1054	switch(uc->cmd) {
1055	case UC_SETSYNC:
1056	case UC_SETTAGS:
1057	case UC_SETWIDE:
1058	case UC_SETFLAG:
1059	case UC_RESETDEV:
1060	case UC_CLEARDEV:
1061		SKIP_SPACES(ptr, len);
1062		if ((arg_len = is_keyword(ptr, len, "all")) != 0) {
1063			ptr += arg_len; len -= arg_len;
1064			uc->target = ~0;
1065		} else {
1066			GET_INT_ARG(ptr, len, target);
1067			uc->target = (1<<target);
1068#ifdef DEBUG_PROC_INFO
1069printk("sym_user_command: target=%ld\n", target);
1070#endif
1071		}
1072		break;
1073	}
1074
1075	switch(uc->cmd) {
1076	case UC_SETVERBOSE:
1077	case UC_SETSYNC:
1078	case UC_SETTAGS:
1079	case UC_SETWIDE:
1080		SKIP_SPACES(ptr, len);
1081		GET_INT_ARG(ptr, len, uc->data);
1082#ifdef DEBUG_PROC_INFO
1083printk("sym_user_command: data=%ld\n", uc->data);
1084#endif
1085		break;
1086#ifdef SYM_LINUX_DEBUG_CONTROL_SUPPORT
1087	case UC_SETDEBUG:
1088		while (len > 0) {
1089			SKIP_SPACES(ptr, len);
1090			if	((arg_len = is_keyword(ptr, len, "alloc")))
1091				uc->data |= DEBUG_ALLOC;
1092			else if	((arg_len = is_keyword(ptr, len, "phase")))
1093				uc->data |= DEBUG_PHASE;
1094			else if	((arg_len = is_keyword(ptr, len, "queue")))
1095				uc->data |= DEBUG_QUEUE;
1096			else if	((arg_len = is_keyword(ptr, len, "result")))
1097				uc->data |= DEBUG_RESULT;
1098			else if	((arg_len = is_keyword(ptr, len, "scatter")))
1099				uc->data |= DEBUG_SCATTER;
1100			else if	((arg_len = is_keyword(ptr, len, "script")))
1101				uc->data |= DEBUG_SCRIPT;
1102			else if	((arg_len = is_keyword(ptr, len, "tiny")))
1103				uc->data |= DEBUG_TINY;
1104			else if	((arg_len = is_keyword(ptr, len, "timing")))
1105				uc->data |= DEBUG_TIMING;
1106			else if	((arg_len = is_keyword(ptr, len, "nego")))
1107				uc->data |= DEBUG_NEGO;
1108			else if	((arg_len = is_keyword(ptr, len, "tags")))
1109				uc->data |= DEBUG_TAGS;
1110			else if	((arg_len = is_keyword(ptr, len, "pointer")))
1111				uc->data |= DEBUG_POINTER;
1112			else
1113				return -EINVAL;
1114			ptr += arg_len; len -= arg_len;
1115		}
1116#ifdef DEBUG_PROC_INFO
1117printk("sym_user_command: data=%ld\n", uc->data);
1118#endif
1119		break;
1120#endif /* SYM_LINUX_DEBUG_CONTROL_SUPPORT */
1121	case UC_SETFLAG:
1122		while (len > 0) {
1123			SKIP_SPACES(ptr, len);
1124			if	((arg_len = is_keyword(ptr, len, "no_disc")))
1125				uc->data &= ~SYM_DISC_ENABLED;
1126			else
1127				return -EINVAL;
1128			ptr += arg_len; len -= arg_len;
1129		}
1130		break;
1131	default:
1132		break;
1133	}
1134
1135	if (len)
1136		return -EINVAL;
1137	else {
1138		unsigned long flags;
1139
1140		spin_lock_irqsave(shost->host_lock, flags);
1141		sym_exec_user_command(np, uc);
1142		spin_unlock_irqrestore(shost->host_lock, flags);
1143	}
1144	return length;
1145}
1146
1147#endif	/* SYM_LINUX_USER_COMMAND_SUPPORT */
1148
1149
1150/*
1151 *  Copy formatted information into the input buffer.
1152 */
1153static int sym_show_info(struct seq_file *m, struct Scsi_Host *shost)
1154{
1155#ifdef SYM_LINUX_USER_INFO_SUPPORT
1156	struct sym_data *sym_data = shost_priv(shost);
1157	struct pci_dev *pdev = sym_data->pdev;
1158	struct sym_hcb *np = sym_data->ncb;
1159
1160	seq_printf(m, "Chip " NAME53C "%s, device id 0x%x, "
1161		 "revision id 0x%x\n", np->s.chip_name,
1162		 pdev->device, pdev->revision);
1163	seq_printf(m, "At PCI address %s, IRQ %u\n",
1164			 pci_name(pdev), pdev->irq);
1165	seq_printf(m, "Min. period factor %d, %s SCSI BUS%s\n",
1166		 (int) (np->minsync_dt ? np->minsync_dt : np->minsync),
1167		 np->maxwide ? "Wide" : "Narrow",
1168		 np->minsync_dt ? ", DT capable" : "");
1169
1170	seq_printf(m, "Max. started commands %d, "
1171		 "max. commands per LUN %d\n",
1172		 SYM_CONF_MAX_START, SYM_CONF_MAX_TAG);
1173
1174	return 0;
1175#else
1176	return -EINVAL;
1177#endif /* SYM_LINUX_USER_INFO_SUPPORT */
1178}
1179
1180#endif /* SYM_LINUX_PROC_INFO_SUPPORT */
1181
1182/*
1183 * Free resources claimed by sym_iomap_device().  Note that
1184 * sym_free_resources() should be used instead of this function after calling
1185 * sym_attach().
1186 */
1187static void sym_iounmap_device(struct sym_device *device)
1188{
1189	if (device->s.ioaddr)
1190		pci_iounmap(device->pdev, device->s.ioaddr);
1191	if (device->s.ramaddr)
1192		pci_iounmap(device->pdev, device->s.ramaddr);
1193}
1194
1195/*
1196 *	Free controller resources.
1197 */
1198static void sym_free_resources(struct sym_hcb *np, struct pci_dev *pdev,
1199		int do_free_irq)
1200{
1201	/*
1202	 *  Free O/S specific resources.
1203	 */
1204	if (do_free_irq)
1205		free_irq(pdev->irq, np->s.host);
1206	if (np->s.ioaddr)
1207		pci_iounmap(pdev, np->s.ioaddr);
1208	if (np->s.ramaddr)
1209		pci_iounmap(pdev, np->s.ramaddr);
1210	/*
1211	 *  Free O/S independent resources.
1212	 */
1213	sym_hcb_free(np);
1214
1215	sym_mfree_dma(np, sizeof(*np), "HCB");
1216}
1217
1218/*
1219 *  Host attach and initialisations.
1220 *
1221 *  Allocate host data and ncb structure.
1222 *  Remap MMIO region.
1223 *  Do chip initialization.
1224 *  If all is OK, install interrupt handling and
1225 *  start the timer daemon.
1226 */
1227static struct Scsi_Host *sym_attach(const struct scsi_host_template *tpnt, int unit,
1228				    struct sym_device *dev)
1229{
1230	struct sym_data *sym_data;
1231	struct sym_hcb *np = NULL;
1232	struct Scsi_Host *shost = NULL;
1233	struct pci_dev *pdev = dev->pdev;
1234	unsigned long flags;
1235	struct sym_fw *fw;
1236	int do_free_irq = 0;
1237
1238	printk(KERN_INFO "sym%d: <%s> rev 0x%x at pci %s irq %u\n",
1239		unit, dev->chip.name, pdev->revision, pci_name(pdev),
1240		pdev->irq);
1241
1242	/*
1243	 *  Get the firmware for this chip.
1244	 */
1245	fw = sym_find_firmware(&dev->chip);
1246	if (!fw)
1247		goto attach_failed;
1248
1249	shost = scsi_host_alloc(tpnt, sizeof(*sym_data));
1250	if (!shost)
1251		goto attach_failed;
1252	sym_data = shost_priv(shost);
1253
1254	/*
1255	 *  Allocate immediately the host control block,
1256	 *  since we are only expecting to succeed. :)
1257	 *  We keep track in the HCB of all the resources that
1258	 *  are to be released on error.
1259	 */
1260	np = __sym_calloc_dma(&pdev->dev, sizeof(*np), "HCB");
1261	if (!np)
1262		goto attach_failed;
1263	np->bus_dmat = &pdev->dev; /* Result in 1 DMA pool per HBA */
1264	sym_data->ncb = np;
1265	sym_data->pdev = pdev;
1266	np->s.host = shost;
1267
1268	pci_set_drvdata(pdev, shost);
1269
1270	/*
1271	 *  Copy some useful infos to the HCB.
1272	 */
1273	np->hcb_ba	= vtobus(np);
1274	np->verbose	= sym_driver_setup.verbose;
1275	np->s.unit	= unit;
1276	np->features	= dev->chip.features;
1277	np->clock_divn	= dev->chip.nr_divisor;
1278	np->maxoffs	= dev->chip.offset_max;
1279	np->maxburst	= dev->chip.burst_max;
1280	np->myaddr	= dev->host_id;
1281	np->mmio_ba	= (u32)dev->mmio_base;
1282	np->ram_ba	= (u32)dev->ram_base;
1283	np->s.ioaddr	= dev->s.ioaddr;
1284	np->s.ramaddr	= dev->s.ramaddr;
1285
1286	/*
1287	 *  Edit its name.
1288	 */
1289	strscpy(np->s.chip_name, dev->chip.name, sizeof(np->s.chip_name));
1290	sprintf(np->s.inst_name, "sym%d", np->s.unit);
1291
1292	if ((SYM_CONF_DMA_ADDRESSING_MODE > 0) && (np->features & FE_DAC) &&
1293			!dma_set_mask(&pdev->dev, DMA_DAC_MASK)) {
1294		set_dac(np);
1295	} else if (dma_set_mask(&pdev->dev, DMA_BIT_MASK(32))) {
1296		printf_warning("%s: No suitable DMA available\n", sym_name(np));
1297		goto attach_failed;
1298	}
1299
1300	if (sym_hcb_attach(shost, fw, dev->nvram))
1301		goto attach_failed;
1302
1303	/*
1304	 *  Install the interrupt handler.
1305	 *  If we synchonize the C code with SCRIPTS on interrupt,
1306	 *  we do not want to share the INTR line at all.
1307	 */
1308	if (request_irq(pdev->irq, sym53c8xx_intr, IRQF_SHARED, NAME53C8XX,
1309			shost)) {
1310		printf_err("%s: request irq %u failure\n",
1311			sym_name(np), pdev->irq);
1312		goto attach_failed;
1313	}
1314	do_free_irq = 1;
1315
1316	/*
1317	 *  After SCSI devices have been opened, we cannot
1318	 *  reset the bus safely, so we do it here.
1319	 */
1320	spin_lock_irqsave(shost->host_lock, flags);
1321	if (sym_reset_scsi_bus(np, 0))
1322		goto reset_failed;
1323
1324	/*
1325	 *  Start the SCRIPTS.
1326	 */
1327	sym_start_up(shost, 1);
1328
1329	/*
1330	 *  Start the timer daemon
1331	 */
1332	timer_setup(&np->s.timer, sym53c8xx_timer, 0);
1333	np->s.lasttime=0;
1334	sym_timer (np);
1335
1336	/*
1337	 *  Fill Linux host instance structure
1338	 *  and return success.
1339	 */
1340	shost->max_channel	= 0;
1341	shost->this_id		= np->myaddr;
1342	shost->max_id		= np->maxwide ? 16 : 8;
1343	shost->max_lun		= SYM_CONF_MAX_LUN;
1344	shost->unique_id	= pci_resource_start(pdev, 0);
1345	shost->cmd_per_lun	= SYM_CONF_MAX_TAG;
1346	shost->can_queue	= (SYM_CONF_MAX_START-2);
1347	shost->sg_tablesize	= SYM_CONF_MAX_SG;
1348	shost->max_cmd_len	= 16;
1349	BUG_ON(sym2_transport_template == NULL);
1350	shost->transportt	= sym2_transport_template;
1351
1352	/* 53c896 rev 1 errata: DMA may not cross 16MB boundary */
1353	if (pdev->device == PCI_DEVICE_ID_NCR_53C896 && pdev->revision < 2)
1354		shost->dma_boundary = 0xFFFFFF;
1355
1356	spin_unlock_irqrestore(shost->host_lock, flags);
1357
1358	return shost;
1359
1360 reset_failed:
1361	printf_err("%s: FATAL ERROR: CHECK SCSI BUS - CABLES, "
1362		   "TERMINATION, DEVICE POWER etc.!\n", sym_name(np));
1363	spin_unlock_irqrestore(shost->host_lock, flags);
1364 attach_failed:
1365	printf_info("sym%d: giving up ...\n", unit);
1366	if (np)
1367		sym_free_resources(np, pdev, do_free_irq);
1368	else
1369		sym_iounmap_device(dev);
1370	if (shost)
1371		scsi_host_put(shost);
1372
1373	return NULL;
1374}
1375
1376
1377/*
1378 *    Detect and try to read SYMBIOS and TEKRAM NVRAM.
1379 */
1380#if SYM_CONF_NVRAM_SUPPORT
1381static void sym_get_nvram(struct sym_device *devp, struct sym_nvram *nvp)
1382{
1383	devp->nvram = nvp;
1384	nvp->type = 0;
1385
1386	sym_read_nvram(devp, nvp);
1387}
1388#else
1389static inline void sym_get_nvram(struct sym_device *devp, struct sym_nvram *nvp)
1390{
1391}
1392#endif	/* SYM_CONF_NVRAM_SUPPORT */
1393
1394static int sym_check_supported(struct sym_device *device)
1395{
1396	struct sym_chip *chip;
1397	struct pci_dev *pdev = device->pdev;
1398	unsigned long io_port = pci_resource_start(pdev, 0);
1399	int i;
1400
1401	/*
1402	 *  If user excluded this chip, do not initialize it.
1403	 *  I hate this code so much.  Must kill it.
1404	 */
1405	if (io_port) {
1406		for (i = 0 ; i < 8 ; i++) {
1407			if (sym_driver_setup.excludes[i] == io_port)
1408				return -ENODEV;
1409		}
1410	}
1411
1412	/*
1413	 * Check if the chip is supported.  Then copy the chip description
1414	 * to our device structure so we can make it match the actual device
1415	 * and options.
1416	 */
1417	chip = sym_lookup_chip_table(pdev->device, pdev->revision);
1418	if (!chip) {
1419		dev_info(&pdev->dev, "device not supported\n");
1420		return -ENODEV;
1421	}
1422	memcpy(&device->chip, chip, sizeof(device->chip));
1423
1424	return 0;
1425}
1426
1427/*
1428 * Ignore Symbios chips controlled by various RAID controllers.
1429 * These controllers set value 0x52414944 at RAM end - 16.
1430 */
1431static int sym_check_raid(struct sym_device *device)
1432{
1433	unsigned int ram_size, ram_val;
1434
1435	if (!device->s.ramaddr)
1436		return 0;
1437
1438	if (device->chip.features & FE_RAM8K)
1439		ram_size = 8192;
1440	else
1441		ram_size = 4096;
1442
1443	ram_val = readl(device->s.ramaddr + ram_size - 16);
1444	if (ram_val != 0x52414944)
1445		return 0;
1446
1447	dev_info(&device->pdev->dev,
1448			"not initializing, driven by RAID controller.\n");
1449	return -ENODEV;
1450}
1451
1452static int sym_set_workarounds(struct sym_device *device)
1453{
1454	struct sym_chip *chip = &device->chip;
1455	struct pci_dev *pdev = device->pdev;
1456	u_short status_reg;
1457
1458	/*
1459	 *  (ITEM 12 of a DEL about the 896 I haven't yet).
1460	 *  We must ensure the chip will use WRITE AND INVALIDATE.
1461	 *  The revision number limit is for now arbitrary.
1462	 */
1463	if (pdev->device == PCI_DEVICE_ID_NCR_53C896 && pdev->revision < 0x4) {
1464		chip->features	|= (FE_WRIE | FE_CLSE);
1465	}
1466
1467	/* If the chip can do Memory Write Invalidate, enable it */
1468	if (chip->features & FE_WRIE) {
1469		if (pci_set_mwi(pdev))
1470			return -ENODEV;
1471	}
1472
1473	/*
1474	 *  Work around for errant bit in 895A. The 66Mhz
1475	 *  capable bit is set erroneously. Clear this bit.
1476	 *  (Item 1 DEL 533)
1477	 *
1478	 *  Make sure Config space and Features agree.
1479	 *
1480	 *  Recall: writes are not normal to status register -
1481	 *  write a 1 to clear and a 0 to leave unchanged.
1482	 *  Can only reset bits.
1483	 */
1484	pci_read_config_word(pdev, PCI_STATUS, &status_reg);
1485	if (chip->features & FE_66MHZ) {
1486		if (!(status_reg & PCI_STATUS_66MHZ))
1487			chip->features &= ~FE_66MHZ;
1488	} else {
1489		if (status_reg & PCI_STATUS_66MHZ) {
1490			status_reg = PCI_STATUS_66MHZ;
1491			pci_write_config_word(pdev, PCI_STATUS, status_reg);
1492			pci_read_config_word(pdev, PCI_STATUS, &status_reg);
1493		}
1494	}
1495
1496	return 0;
1497}
1498
1499/*
1500 * Map HBA registers and on-chip SRAM (if present).
1501 */
1502static int sym_iomap_device(struct sym_device *device)
1503{
1504	struct pci_dev *pdev = device->pdev;
1505	struct pci_bus_region bus_addr;
1506	int i = 2;
1507
1508	pcibios_resource_to_bus(pdev->bus, &bus_addr, &pdev->resource[1]);
1509	device->mmio_base = bus_addr.start;
1510
1511	if (device->chip.features & FE_RAM) {
1512		/*
1513		 * If the BAR is 64-bit, resource 2 will be occupied by the
1514		 * upper 32 bits
1515		 */
1516		if (!pdev->resource[i].flags)
1517			i++;
1518		pcibios_resource_to_bus(pdev->bus, &bus_addr,
1519					&pdev->resource[i]);
1520		device->ram_base = bus_addr.start;
1521	}
1522
1523#ifdef CONFIG_SCSI_SYM53C8XX_MMIO
1524	if (device->mmio_base)
1525		device->s.ioaddr = pci_iomap(pdev, 1,
1526						pci_resource_len(pdev, 1));
1527#endif
1528	if (!device->s.ioaddr)
1529		device->s.ioaddr = pci_iomap(pdev, 0,
1530						pci_resource_len(pdev, 0));
1531	if (!device->s.ioaddr) {
1532		dev_err(&pdev->dev, "could not map registers; giving up.\n");
1533		return -EIO;
1534	}
1535	if (device->ram_base) {
1536		device->s.ramaddr = pci_iomap(pdev, i,
1537						pci_resource_len(pdev, i));
1538		if (!device->s.ramaddr) {
1539			dev_warn(&pdev->dev,
1540				"could not map SRAM; continuing anyway.\n");
1541			device->ram_base = 0;
1542		}
1543	}
1544
1545	return 0;
1546}
1547
1548/*
1549 * The NCR PQS and PDS cards are constructed as a DEC bridge
1550 * behind which sits a proprietary NCR memory controller and
1551 * either four or two 53c875s as separate devices.  We can tell
1552 * if an 875 is part of a PQS/PDS or not since if it is, it will
1553 * be on the same bus as the memory controller.  In its usual
1554 * mode of operation, the 875s are slaved to the memory
1555 * controller for all transfers.  To operate with the Linux
1556 * driver, the memory controller is disabled and the 875s
1557 * freed to function independently.  The only wrinkle is that
1558 * the preset SCSI ID (which may be zero) must be read in from
1559 * a special configuration space register of the 875.
1560 */
1561static void sym_config_pqs(struct pci_dev *pdev, struct sym_device *sym_dev)
1562{
1563	int slot;
1564	u8 tmp;
1565
1566	for (slot = 0; slot < 256; slot++) {
1567		struct pci_dev *memc = pci_get_slot(pdev->bus, slot);
1568
1569		if (!memc || memc->vendor != 0x101a || memc->device == 0x0009) {
1570			pci_dev_put(memc);
1571			continue;
1572		}
1573
1574		/* bit 1: allow individual 875 configuration */
1575		pci_read_config_byte(memc, 0x44, &tmp);
1576		if ((tmp & 0x2) == 0) {
1577			tmp |= 0x2;
1578			pci_write_config_byte(memc, 0x44, tmp);
1579		}
1580
1581		/* bit 2: drive individual 875 interrupts to the bus */
1582		pci_read_config_byte(memc, 0x45, &tmp);
1583		if ((tmp & 0x4) == 0) {
1584			tmp |= 0x4;
1585			pci_write_config_byte(memc, 0x45, tmp);
1586		}
1587
1588		pci_dev_put(memc);
1589		break;
1590	}
1591
1592	pci_read_config_byte(pdev, 0x84, &tmp);
1593	sym_dev->host_id = tmp;
1594}
1595
1596/*
1597 *  Called before unloading the module.
1598 *  Detach the host.
1599 *  We have to free resources and halt the NCR chip.
1600 */
1601static int sym_detach(struct Scsi_Host *shost, struct pci_dev *pdev)
1602{
1603	struct sym_hcb *np = sym_get_hcb(shost);
1604	printk("%s: detaching ...\n", sym_name(np));
1605
1606	del_timer_sync(&np->s.timer);
1607
1608	/*
1609	 * Reset NCR chip.
1610	 * We should use sym_soft_reset(), but we don't want to do
1611	 * so, since we may not be safe if interrupts occur.
1612	 */
1613	printk("%s: resetting chip\n", sym_name(np));
1614	OUTB(np, nc_istat, SRST);
1615	INB(np, nc_mbox1);
1616	udelay(10);
1617	OUTB(np, nc_istat, 0);
1618
1619	sym_free_resources(np, pdev, 1);
1620	scsi_host_put(shost);
1621
1622	return 1;
1623}
1624
1625/*
1626 * Driver host template.
1627 */
1628static const struct scsi_host_template sym2_template = {
1629	.module			= THIS_MODULE,
1630	.name			= "sym53c8xx",
1631	.info			= sym53c8xx_info,
1632	.cmd_size		= sizeof(struct sym_ucmd),
1633	.queuecommand		= sym53c8xx_queue_command,
1634	.slave_alloc		= sym53c8xx_slave_alloc,
1635	.slave_configure	= sym53c8xx_slave_configure,
1636	.slave_destroy		= sym53c8xx_slave_destroy,
1637	.eh_abort_handler	= sym53c8xx_eh_abort_handler,
1638	.eh_device_reset_handler = sym53c8xx_eh_device_reset_handler,
1639	.eh_bus_reset_handler	= sym53c8xx_eh_bus_reset_handler,
1640	.eh_host_reset_handler	= sym53c8xx_eh_host_reset_handler,
1641	.this_id		= 7,
1642	.max_sectors		= 0xFFFF,
1643#ifdef SYM_LINUX_PROC_INFO_SUPPORT
1644	.show_info		= sym_show_info,
1645#ifdef	SYM_LINUX_USER_COMMAND_SUPPORT
1646	.write_info		= sym_user_command,
1647#endif
1648	.proc_name		= NAME53C8XX,
1649#endif
1650};
1651
1652static int attach_count;
1653
1654static int sym2_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
1655{
1656	struct sym_device sym_dev;
1657	struct sym_nvram nvram;
1658	struct Scsi_Host *shost;
1659	int do_iounmap = 0;
1660	int do_disable_device = 1;
1661
1662	memset(&sym_dev, 0, sizeof(sym_dev));
1663	memset(&nvram, 0, sizeof(nvram));
1664	sym_dev.pdev = pdev;
1665	sym_dev.host_id = SYM_SETUP_HOST_ID;
1666
1667	if (pci_enable_device(pdev))
1668		goto leave;
1669
1670	pci_set_master(pdev);
1671
1672	if (pci_request_regions(pdev, NAME53C8XX))
1673		goto disable;
1674
1675	if (sym_check_supported(&sym_dev))
1676		goto free;
1677
1678	if (sym_iomap_device(&sym_dev))
1679		goto free;
1680	do_iounmap = 1;
1681
1682	if (sym_check_raid(&sym_dev)) {
1683		do_disable_device = 0;	/* Don't disable the device */
1684		goto free;
1685	}
1686
1687	if (sym_set_workarounds(&sym_dev))
1688		goto free;
1689
1690	sym_config_pqs(pdev, &sym_dev);
1691
1692	sym_get_nvram(&sym_dev, &nvram);
1693
1694	do_iounmap = 0; /* Don't sym_iounmap_device() after sym_attach(). */
1695	shost = sym_attach(&sym2_template, attach_count, &sym_dev);
1696	if (!shost)
1697		goto free;
1698
1699	if (scsi_add_host(shost, &pdev->dev))
1700		goto detach;
1701	scsi_scan_host(shost);
1702
1703	attach_count++;
1704
1705	return 0;
1706
1707 detach:
1708	sym_detach(pci_get_drvdata(pdev), pdev);
1709 free:
1710	if (do_iounmap)
1711		sym_iounmap_device(&sym_dev);
1712	pci_release_regions(pdev);
1713 disable:
1714	if (do_disable_device)
1715		pci_disable_device(pdev);
1716 leave:
1717	return -ENODEV;
1718}
1719
1720static void sym2_remove(struct pci_dev *pdev)
1721{
1722	struct Scsi_Host *shost = pci_get_drvdata(pdev);
1723
1724	scsi_remove_host(shost);
1725	sym_detach(shost, pdev);
1726	pci_release_regions(pdev);
1727	pci_disable_device(pdev);
1728
1729	attach_count--;
1730}
1731
1732/**
1733 * sym2_io_error_detected() - called when PCI error is detected
1734 * @pdev: pointer to PCI device
1735 * @state: current state of the PCI slot
1736 */
1737static pci_ers_result_t sym2_io_error_detected(struct pci_dev *pdev,
1738                                         pci_channel_state_t state)
1739{
1740	/* If slot is permanently frozen, turn everything off */
1741	if (state == pci_channel_io_perm_failure) {
1742		sym2_remove(pdev);
1743		return PCI_ERS_RESULT_DISCONNECT;
1744	}
1745
1746	disable_irq(pdev->irq);
1747	pci_disable_device(pdev);
1748
1749	/* Request that MMIO be enabled, so register dump can be taken. */
1750	return PCI_ERS_RESULT_CAN_RECOVER;
1751}
1752
1753/**
1754 * sym2_io_slot_dump - Enable MMIO and dump debug registers
1755 * @pdev: pointer to PCI device
1756 */
1757static pci_ers_result_t sym2_io_slot_dump(struct pci_dev *pdev)
1758{
1759	struct Scsi_Host *shost = pci_get_drvdata(pdev);
1760
1761	sym_dump_registers(shost);
1762
1763	/* Request a slot reset. */
1764	return PCI_ERS_RESULT_NEED_RESET;
1765}
1766
1767/**
1768 * sym2_reset_workarounds - hardware-specific work-arounds
1769 * @pdev: pointer to PCI device
1770 *
1771 * This routine is similar to sym_set_workarounds(), except
1772 * that, at this point, we already know that the device was
1773 * successfully initialized at least once before, and so most
1774 * of the steps taken there are un-needed here.
1775 */
1776static void sym2_reset_workarounds(struct pci_dev *pdev)
1777{
1778	u_short status_reg;
1779	struct sym_chip *chip;
1780
1781	chip = sym_lookup_chip_table(pdev->device, pdev->revision);
1782
1783	/* Work around for errant bit in 895A, in a fashion
1784	 * similar to what is done in sym_set_workarounds().
1785	 */
1786	pci_read_config_word(pdev, PCI_STATUS, &status_reg);
1787	if (!(chip->features & FE_66MHZ) && (status_reg & PCI_STATUS_66MHZ)) {
1788		status_reg = PCI_STATUS_66MHZ;
1789		pci_write_config_word(pdev, PCI_STATUS, status_reg);
1790		pci_read_config_word(pdev, PCI_STATUS, &status_reg);
1791	}
1792}
1793
1794/**
1795 * sym2_io_slot_reset() - called when the pci bus has been reset.
1796 * @pdev: pointer to PCI device
1797 *
1798 * Restart the card from scratch.
1799 */
1800static pci_ers_result_t sym2_io_slot_reset(struct pci_dev *pdev)
1801{
1802	struct Scsi_Host *shost = pci_get_drvdata(pdev);
1803	struct sym_hcb *np = sym_get_hcb(shost);
1804
1805	printk(KERN_INFO "%s: recovering from a PCI slot reset\n",
1806	          sym_name(np));
1807
1808	if (pci_enable_device(pdev)) {
1809		printk(KERN_ERR "%s: Unable to enable after PCI reset\n",
1810		        sym_name(np));
1811		return PCI_ERS_RESULT_DISCONNECT;
1812	}
1813
1814	pci_set_master(pdev);
1815	enable_irq(pdev->irq);
1816
1817	/* If the chip can do Memory Write Invalidate, enable it */
1818	if (np->features & FE_WRIE) {
1819		if (pci_set_mwi(pdev))
1820			return PCI_ERS_RESULT_DISCONNECT;
1821	}
1822
1823	/* Perform work-arounds, analogous to sym_set_workarounds() */
1824	sym2_reset_workarounds(pdev);
1825
1826	/* Perform host reset only on one instance of the card */
1827	if (PCI_FUNC(pdev->devfn) == 0) {
1828		if (sym_reset_scsi_bus(np, 0)) {
1829			printk(KERN_ERR "%s: Unable to reset scsi host\n",
1830			        sym_name(np));
1831			return PCI_ERS_RESULT_DISCONNECT;
1832		}
1833		sym_start_up(shost, 1);
1834	}
1835
1836	return PCI_ERS_RESULT_RECOVERED;
1837}
1838
1839/**
1840 * sym2_io_resume() - resume normal ops after PCI reset
1841 * @pdev: pointer to PCI device
1842 *
1843 * Called when the error recovery driver tells us that its
1844 * OK to resume normal operation. Use completion to allow
1845 * halted scsi ops to resume.
1846 */
1847static void sym2_io_resume(struct pci_dev *pdev)
1848{
1849	struct Scsi_Host *shost = pci_get_drvdata(pdev);
1850	struct sym_data *sym_data = shost_priv(shost);
1851
1852	spin_lock_irq(shost->host_lock);
1853	if (sym_data->io_reset)
1854		complete(sym_data->io_reset);
1855	spin_unlock_irq(shost->host_lock);
1856}
1857
1858static void sym2_get_signalling(struct Scsi_Host *shost)
1859{
1860	struct sym_hcb *np = sym_get_hcb(shost);
1861	enum spi_signal_type type;
1862
1863	switch (np->scsi_mode) {
1864	case SMODE_SE:
1865		type = SPI_SIGNAL_SE;
1866		break;
1867	case SMODE_LVD:
1868		type = SPI_SIGNAL_LVD;
1869		break;
1870	case SMODE_HVD:
1871		type = SPI_SIGNAL_HVD;
1872		break;
1873	default:
1874		type = SPI_SIGNAL_UNKNOWN;
1875		break;
1876	}
1877	spi_signalling(shost) = type;
1878}
1879
1880static void sym2_set_offset(struct scsi_target *starget, int offset)
1881{
1882	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1883	struct sym_hcb *np = sym_get_hcb(shost);
1884	struct sym_tcb *tp = &np->target[starget->id];
1885
1886	tp->tgoal.offset = offset;
1887	tp->tgoal.check_nego = 1;
1888}
1889
1890static void sym2_set_period(struct scsi_target *starget, int period)
1891{
1892	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1893	struct sym_hcb *np = sym_get_hcb(shost);
1894	struct sym_tcb *tp = &np->target[starget->id];
1895
1896	/* have to have DT for these transfers, but DT will also
1897	 * set width, so check that this is allowed */
1898	if (period <= np->minsync && spi_width(starget))
1899		tp->tgoal.dt = 1;
1900
1901	tp->tgoal.period = period;
1902	tp->tgoal.check_nego = 1;
1903}
1904
1905static void sym2_set_width(struct scsi_target *starget, int width)
1906{
1907	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1908	struct sym_hcb *np = sym_get_hcb(shost);
1909	struct sym_tcb *tp = &np->target[starget->id];
1910
1911	/* It is illegal to have DT set on narrow transfers.  If DT is
1912	 * clear, we must also clear IU and QAS.  */
1913	if (width == 0)
1914		tp->tgoal.iu = tp->tgoal.dt = tp->tgoal.qas = 0;
1915
1916	tp->tgoal.width = width;
1917	tp->tgoal.check_nego = 1;
1918}
1919
1920static void sym2_set_dt(struct scsi_target *starget, int dt)
1921{
1922	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1923	struct sym_hcb *np = sym_get_hcb(shost);
1924	struct sym_tcb *tp = &np->target[starget->id];
1925
1926	/* We must clear QAS and IU if DT is clear */
1927	if (dt)
1928		tp->tgoal.dt = 1;
1929	else
1930		tp->tgoal.iu = tp->tgoal.dt = tp->tgoal.qas = 0;
1931	tp->tgoal.check_nego = 1;
1932}
1933
1934#if 0
1935static void sym2_set_iu(struct scsi_target *starget, int iu)
1936{
1937	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1938	struct sym_hcb *np = sym_get_hcb(shost);
1939	struct sym_tcb *tp = &np->target[starget->id];
1940
1941	if (iu)
1942		tp->tgoal.iu = tp->tgoal.dt = 1;
1943	else
1944		tp->tgoal.iu = 0;
1945	tp->tgoal.check_nego = 1;
1946}
1947
1948static void sym2_set_qas(struct scsi_target *starget, int qas)
1949{
1950	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1951	struct sym_hcb *np = sym_get_hcb(shost);
1952	struct sym_tcb *tp = &np->target[starget->id];
1953
1954	if (qas)
1955		tp->tgoal.dt = tp->tgoal.qas = 1;
1956	else
1957		tp->tgoal.qas = 0;
1958	tp->tgoal.check_nego = 1;
1959}
1960#endif
1961
1962static struct spi_function_template sym2_transport_functions = {
1963	.set_offset	= sym2_set_offset,
1964	.show_offset	= 1,
1965	.set_period	= sym2_set_period,
1966	.show_period	= 1,
1967	.set_width	= sym2_set_width,
1968	.show_width	= 1,
1969	.set_dt		= sym2_set_dt,
1970	.show_dt	= 1,
1971#if 0
1972	.set_iu		= sym2_set_iu,
1973	.show_iu	= 1,
1974	.set_qas	= sym2_set_qas,
1975	.show_qas	= 1,
1976#endif
1977	.get_signalling	= sym2_get_signalling,
1978};
1979
1980static struct pci_device_id sym2_id_table[] = {
1981	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C810,
1982	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
1983	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C820,
1984	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL }, /* new */
1985	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C825,
1986	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
1987	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C815,
1988	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
1989	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_LSI_53C810AP,
1990	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL }, /* new */
1991	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C860,
1992	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
1993	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_LSI_53C1510,
1994	  PCI_ANY_ID, PCI_ANY_ID, PCI_CLASS_STORAGE_SCSI<<8,  0xffff00, 0UL },
1995	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C896,
1996	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
1997	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C895,
1998	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
1999	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C885,
2000	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
2001	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C875,
2002	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
2003	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C1510,
2004	  PCI_ANY_ID, PCI_ANY_ID,  PCI_CLASS_STORAGE_SCSI<<8,  0xffff00, 0UL }, /* new */
2005	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_LSI_53C895A,
2006	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
2007	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_LSI_53C875A,
2008	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
2009	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_LSI_53C1010_33,
2010	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
2011	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_LSI_53C1010_66,
2012	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
2013	{ PCI_VENDOR_ID_LSI_LOGIC, PCI_DEVICE_ID_NCR_53C875J,
2014	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
2015	{ 0, }
2016};
2017
2018MODULE_DEVICE_TABLE(pci, sym2_id_table);
2019
2020static const struct pci_error_handlers sym2_err_handler = {
2021	.error_detected	= sym2_io_error_detected,
2022	.mmio_enabled	= sym2_io_slot_dump,
2023	.slot_reset	= sym2_io_slot_reset,
2024	.resume		= sym2_io_resume,
2025};
2026
2027static struct pci_driver sym2_driver = {
2028	.name		= NAME53C8XX,
2029	.id_table	= sym2_id_table,
2030	.probe		= sym2_probe,
2031	.remove		= sym2_remove,
2032	.err_handler 	= &sym2_err_handler,
2033};
2034
2035static int __init sym2_init(void)
2036{
2037	int error;
2038
2039	sym2_setup_params();
2040	sym2_transport_template = spi_attach_transport(&sym2_transport_functions);
2041	if (!sym2_transport_template)
2042		return -ENODEV;
2043
2044	error = pci_register_driver(&sym2_driver);
2045	if (error)
2046		spi_release_transport(sym2_transport_template);
2047	return error;
2048}
2049
2050static void __exit sym2_exit(void)
2051{
2052	pci_unregister_driver(&sym2_driver);
2053	spi_release_transport(sym2_transport_template);
2054}
2055
2056module_init(sym2_init);
2057module_exit(sym2_exit);
2058