1/* 8390.c: A general NS8390 ethernet driver core for linux. */
2/*
3	Written 1992-94 by Donald Becker.
4
5	Copyright 1993 United States Government as represented by the
6	Director, National Security Agency.
7
8	This software may be used and distributed according to the terms
9	of the GNU General Public License, incorporated herein by reference.
10
11	The author may be reached as becker@scyld.com, or C/O
12	Scyld Computing Corporation
13	410 Severn Ave., Suite 210
14	Annapolis MD 21403
15
16
17  This is the chip-specific code for many 8390-based ethernet adaptors.
18  This is not a complete driver, it must be combined with board-specific
19  code such as ne.c, wd.c, 3c503.c, etc.
20
21  Seeing how at least eight drivers use this code, (not counting the
22  PCMCIA ones either) it is easy to break some card by what seems like
23  a simple innocent change. Please contact me or Donald if you think
24  you have found something that needs changing. -- PG
25
26
27  Changelog:
28
29  Paul Gortmaker	: remove set_bit lock, other cleanups.
30  Paul Gortmaker	: add ei_get_8390_hdr() so we can pass skb's to
31			  ei_block_input() for eth_io_copy_and_sum().
32  Paul Gortmaker	: exchange static int ei_pingpong for a #define,
33			  also add better Tx error handling.
34  Paul Gortmaker	: rewrite Rx overrun handling as per NS specs.
35  Alexey Kuznetsov	: use the 8390's six bit hash multicast filter.
36  Paul Gortmaker	: tweak ANK's above multicast changes a bit.
37  Paul Gortmaker	: update packet statistics for v2.1.x
38  Alan Cox		: support arbitrary stupid port mappings on the
39			  68K Macintosh. Support >16bit I/O spaces
40  Paul Gortmaker	: add kmod support for auto-loading of the 8390
41			  module by all drivers that require it.
42  Alan Cox		: Spinlocking work, added 'BUG_83C690'
43  Paul Gortmaker	: Separate out Tx timeout code from Tx path.
44  Paul Gortmaker	: Remove old unused single Tx buffer code.
45  Hayato Fujiwara	: Add m32r support.
46  Paul Gortmaker	: use skb_padto() instead of stack scratch area
47
48  Sources:
49  The National Semiconductor LAN Databook, and the 3Com 3c503 databook.
50
51  */
52
53#include <linux/build_bug.h>
54#include <linux/module.h>
55#include <linux/kernel.h>
56#include <linux/jiffies.h>
57#include <linux/fs.h>
58#include <linux/types.h>
59#include <linux/string.h>
60#include <linux/bitops.h>
61#include <linux/uaccess.h>
62#include <linux/io.h>
63#include <asm/irq.h>
64#include <linux/delay.h>
65#include <linux/errno.h>
66#include <linux/fcntl.h>
67#include <linux/in.h>
68#include <linux/interrupt.h>
69#include <linux/init.h>
70#include <linux/crc32.h>
71
72#include <linux/netdevice.h>
73#include <linux/etherdevice.h>
74
75#define NS8390_CORE
76#include "8390.h"
77
78#define BUG_83C690
79
80/* These are the operational function interfaces to board-specific
81   routines.
82	void reset_8390(struct net_device *dev)
83		Resets the board associated with DEV, including a hardware reset of
84		the 8390.  This is only called when there is a transmit timeout, and
85		it is always followed by 8390_init().
86	void block_output(struct net_device *dev, int count, const unsigned char *buf,
87					  int start_page)
88		Write the COUNT bytes of BUF to the packet buffer at START_PAGE.  The
89		"page" value uses the 8390's 256-byte pages.
90	void get_8390_hdr(struct net_device *dev, struct e8390_hdr *hdr, int ring_page)
91		Read the 4 byte, page aligned 8390 header. *If* there is a
92		subsequent read, it will be of the rest of the packet.
93	void block_input(struct net_device *dev, int count, struct sk_buff *skb, int ring_offset)
94		Read COUNT bytes from the packet buffer into the skb data area. Start
95		reading from RING_OFFSET, the address as the 8390 sees it.  This will always
96		follow the read of the 8390 header.
97*/
98#define ei_reset_8390 (ei_local->reset_8390)
99#define ei_block_output (ei_local->block_output)
100#define ei_block_input (ei_local->block_input)
101#define ei_get_8390_hdr (ei_local->get_8390_hdr)
102
103/* Index to functions. */
104static void ei_tx_intr(struct net_device *dev);
105static void ei_tx_err(struct net_device *dev);
106static void ei_receive(struct net_device *dev);
107static void ei_rx_overrun(struct net_device *dev);
108
109/* Routines generic to NS8390-based boards. */
110static void NS8390_trigger_send(struct net_device *dev, unsigned int length,
111								int start_page);
112static void do_set_multicast_list(struct net_device *dev);
113static void __NS8390_init(struct net_device *dev, int startp);
114
115static unsigned version_printed;
116static int msg_enable;
117static const int default_msg_level = (NETIF_MSG_DRV | NETIF_MSG_PROBE | NETIF_MSG_RX_ERR |
118				     NETIF_MSG_TX_ERR);
119module_param(msg_enable, int, 0444);
120MODULE_PARM_DESC(msg_enable, "Debug message level (see linux/netdevice.h for bitmap)");
121
122/*
123 *	SMP and the 8390 setup.
124 *
125 *	The 8390 isn't exactly designed to be multithreaded on RX/TX. There is
126 *	a page register that controls bank and packet buffer access. We guard
127 *	this with ei_local->page_lock. Nobody should assume or set the page other
128 *	than zero when the lock is not held. Lock holders must restore page 0
129 *	before unlocking. Even pure readers must take the lock to protect in
130 *	page 0.
131 *
132 *	To make life difficult the chip can also be very slow. We therefore can't
133 *	just use spinlocks. For the longer lockups we disable the irq the device
134 *	sits on and hold the lock. We must hold the lock because there is a dual
135 *	processor case other than interrupts (get stats/set multicast list in
136 *	parallel with each other and transmit).
137 *
138 *	Note: in theory we can just disable the irq on the card _but_ there is
139 *	a latency on SMP irq delivery. So we can easily go "disable irq" "sync irqs"
140 *	enter lock, take the queued irq. So we waddle instead of flying.
141 *
142 *	Finally by special arrangement for the purpose of being generally
143 *	annoying the transmit function is called bh atomic. That places
144 *	restrictions on the user context callers as disable_irq won't save
145 *	them.
146 *
147 *	Additional explanation of problems with locking by Alan Cox:
148 *
149 *	"The author (me) didn't use spin_lock_irqsave because the slowness of the
150 *	card means that approach caused horrible problems like losing serial data
151 *	at 38400 baud on some chips. Remember many 8390 nics on PCI were ISA
152 *	chips with FPGA front ends.
153 *
154 *	Ok the logic behind the 8390 is very simple:
155 *
156 *	Things to know
157 *		- IRQ delivery is asynchronous to the PCI bus
158 *		- Blocking the local CPU IRQ via spin locks was too slow
159 *		- The chip has register windows needing locking work
160 *
161 *	So the path was once (I say once as people appear to have changed it
162 *	in the mean time and it now looks rather bogus if the changes to use
163 *	disable_irq_nosync_irqsave are disabling the local IRQ)
164 *
165 *
166 *		Take the page lock
167 *		Mask the IRQ on chip
168 *		Disable the IRQ (but not mask locally- someone seems to have
169 *			broken this with the lock validator stuff)
170 *			[This must be _nosync as the page lock may otherwise
171 *				deadlock us]
172 *		Drop the page lock and turn IRQs back on
173 *
174 *		At this point an existing IRQ may still be running but we can't
175 *		get a new one
176 *
177 *		Take the lock (so we know the IRQ has terminated) but don't mask
178 *	the IRQs on the processor
179 *		Set irqlock [for debug]
180 *
181 *		Transmit (slow as ****)
182 *
183 *		re-enable the IRQ
184 *
185 *
186 *	We have to use disable_irq because otherwise you will get delayed
187 *	interrupts on the APIC bus deadlocking the transmit path.
188 *
189 *	Quite hairy but the chip simply wasn't designed for SMP and you can't
190 *	even ACK an interrupt without risking corrupting other parallel
191 *	activities on the chip." [lkml, 25 Jul 2007]
192 */
193
194
195
196/**
197 * ei_open - Open/initialize the board.
198 * @dev: network device to initialize
199 *
200 * This routine goes all-out, setting everything
201 * up anew at each open, even though many of these registers should only
202 * need to be set once at boot.
203 */
204static int __ei_open(struct net_device *dev)
205{
206	unsigned long flags;
207	struct ei_device *ei_local = netdev_priv(dev);
208
209	if (dev->watchdog_timeo <= 0)
210		dev->watchdog_timeo = TX_TIMEOUT;
211
212	/*
213	 *	Grab the page lock so we own the register set, then call
214	 *	the init function.
215	 */
216
217	spin_lock_irqsave(&ei_local->page_lock, flags);
218	__NS8390_init(dev, 1);
219	/* Set the flag before we drop the lock, That way the IRQ arrives
220	   after its set and we get no silly warnings */
221	netif_start_queue(dev);
222	spin_unlock_irqrestore(&ei_local->page_lock, flags);
223	ei_local->irqlock = 0;
224	return 0;
225}
226
227/**
228 * ei_close - shut down network device
229 * @dev: network device to close
230 *
231 * Opposite of ei_open(). Only used when "ifconfig <devname> down" is done.
232 */
233static int __ei_close(struct net_device *dev)
234{
235	struct ei_device *ei_local = netdev_priv(dev);
236	unsigned long flags;
237
238	/*
239	 *	Hold the page lock during close
240	 */
241
242	spin_lock_irqsave(&ei_local->page_lock, flags);
243	__NS8390_init(dev, 0);
244	spin_unlock_irqrestore(&ei_local->page_lock, flags);
245	netif_stop_queue(dev);
246	return 0;
247}
248
249/**
250 * ei_tx_timeout - handle transmit time out condition
251 * @dev: network device which has apparently fallen asleep
252 *
253 * Called by kernel when device never acknowledges a transmit has
254 * completed (or failed) - i.e. never posted a Tx related interrupt.
255 */
256
257static void __ei_tx_timeout(struct net_device *dev, unsigned int txqueue)
258{
259	unsigned long e8390_base = dev->base_addr;
260	struct ei_device *ei_local = netdev_priv(dev);
261	int txsr, isr, tickssofar = jiffies - dev_trans_start(dev);
262	unsigned long flags;
263
264	dev->stats.tx_errors++;
265
266	spin_lock_irqsave(&ei_local->page_lock, flags);
267	txsr = ei_inb(e8390_base+EN0_TSR);
268	isr = ei_inb(e8390_base+EN0_ISR);
269	spin_unlock_irqrestore(&ei_local->page_lock, flags);
270
271	netdev_dbg(dev, "Tx timed out, %s TSR=%#2x, ISR=%#2x, t=%d\n",
272		   (txsr & ENTSR_ABT) ? "excess collisions." :
273		   (isr) ? "lost interrupt?" : "cable problem?",
274		   txsr, isr, tickssofar);
275
276	if (!isr && !dev->stats.tx_packets) {
277		/* The 8390 probably hasn't gotten on the cable yet. */
278		ei_local->interface_num ^= 1;   /* Try a different xcvr.  */
279	}
280
281	/* Ugly but a reset can be slow, yet must be protected */
282
283	disable_irq_nosync_lockdep(dev->irq);
284	spin_lock(&ei_local->page_lock);
285
286	/* Try to restart the card.  Perhaps the user has fixed something. */
287	ei_reset_8390(dev);
288	__NS8390_init(dev, 1);
289
290	spin_unlock(&ei_local->page_lock);
291	enable_irq_lockdep(dev->irq);
292	netif_wake_queue(dev);
293}
294
295/**
296 * ei_start_xmit - begin packet transmission
297 * @skb: packet to be sent
298 * @dev: network device to which packet is sent
299 *
300 * Sends a packet to an 8390 network device.
301 */
302
303static netdev_tx_t __ei_start_xmit(struct sk_buff *skb,
304				   struct net_device *dev)
305{
306	unsigned long e8390_base = dev->base_addr;
307	struct ei_device *ei_local = netdev_priv(dev);
308	int send_length = skb->len, output_page;
309	unsigned long flags;
310	char buf[ETH_ZLEN];
311	char *data = skb->data;
312
313	if (skb->len < ETH_ZLEN) {
314		memset(buf, 0, ETH_ZLEN);	/* more efficient than doing just the needed bits */
315		memcpy(buf, data, skb->len);
316		send_length = ETH_ZLEN;
317		data = buf;
318	}
319
320	/* Mask interrupts from the ethercard.
321	   SMP: We have to grab the lock here otherwise the IRQ handler
322	   on another CPU can flip window and race the IRQ mask set. We end
323	   up trashing the mcast filter not disabling irqs if we don't lock */
324
325	spin_lock_irqsave(&ei_local->page_lock, flags);
326	ei_outb_p(0x00, e8390_base + EN0_IMR);
327	spin_unlock_irqrestore(&ei_local->page_lock, flags);
328
329
330	/*
331	 *	Slow phase with lock held.
332	 */
333
334	disable_irq_nosync_lockdep_irqsave(dev->irq, &flags);
335
336	spin_lock(&ei_local->page_lock);
337
338	ei_local->irqlock = 1;
339
340	/*
341	 * We have two Tx slots available for use. Find the first free
342	 * slot, and then perform some sanity checks. With two Tx bufs,
343	 * you get very close to transmitting back-to-back packets. With
344	 * only one Tx buf, the transmitter sits idle while you reload the
345	 * card, leaving a substantial gap between each transmitted packet.
346	 */
347
348	if (ei_local->tx1 == 0) {
349		output_page = ei_local->tx_start_page;
350		ei_local->tx1 = send_length;
351		if ((netif_msg_tx_queued(ei_local)) &&
352		    ei_local->tx2 > 0)
353			netdev_dbg(dev,
354				   "idle transmitter tx2=%d, lasttx=%d, txing=%d\n",
355				   ei_local->tx2, ei_local->lasttx, ei_local->txing);
356	} else if (ei_local->tx2 == 0) {
357		output_page = ei_local->tx_start_page + TX_PAGES/2;
358		ei_local->tx2 = send_length;
359		if ((netif_msg_tx_queued(ei_local)) &&
360		    ei_local->tx1 > 0)
361			netdev_dbg(dev,
362				   "idle transmitter, tx1=%d, lasttx=%d, txing=%d\n",
363				   ei_local->tx1, ei_local->lasttx, ei_local->txing);
364	} else {			/* We should never get here. */
365		netif_dbg(ei_local, tx_err, dev,
366			  "No Tx buffers free! tx1=%d tx2=%d last=%d\n",
367			  ei_local->tx1, ei_local->tx2, ei_local->lasttx);
368		ei_local->irqlock = 0;
369		netif_stop_queue(dev);
370		ei_outb_p(ENISR_ALL, e8390_base + EN0_IMR);
371		spin_unlock(&ei_local->page_lock);
372		enable_irq_lockdep_irqrestore(dev->irq, &flags);
373		dev->stats.tx_errors++;
374		return NETDEV_TX_BUSY;
375	}
376
377	/*
378	 * Okay, now upload the packet and trigger a send if the transmitter
379	 * isn't already sending. If it is busy, the interrupt handler will
380	 * trigger the send later, upon receiving a Tx done interrupt.
381	 */
382
383	ei_block_output(dev, send_length, data, output_page);
384
385	if (!ei_local->txing) {
386		ei_local->txing = 1;
387		NS8390_trigger_send(dev, send_length, output_page);
388		if (output_page == ei_local->tx_start_page) {
389			ei_local->tx1 = -1;
390			ei_local->lasttx = -1;
391		} else {
392			ei_local->tx2 = -1;
393			ei_local->lasttx = -2;
394		}
395	} else
396		ei_local->txqueue++;
397
398	if (ei_local->tx1 && ei_local->tx2)
399		netif_stop_queue(dev);
400	else
401		netif_start_queue(dev);
402
403	/* Turn 8390 interrupts back on. */
404	ei_local->irqlock = 0;
405	ei_outb_p(ENISR_ALL, e8390_base + EN0_IMR);
406
407	spin_unlock(&ei_local->page_lock);
408	enable_irq_lockdep_irqrestore(dev->irq, &flags);
409	skb_tx_timestamp(skb);
410	dev_consume_skb_any(skb);
411	dev->stats.tx_bytes += send_length;
412
413	return NETDEV_TX_OK;
414}
415
416/**
417 * ei_interrupt - handle the interrupts from an 8390
418 * @irq: interrupt number
419 * @dev_id: a pointer to the net_device
420 *
421 * Handle the ether interface interrupts. We pull packets from
422 * the 8390 via the card specific functions and fire them at the networking
423 * stack. We also handle transmit completions and wake the transmit path if
424 * necessary. We also update the counters and do other housekeeping as
425 * needed.
426 */
427
428static irqreturn_t __ei_interrupt(int irq, void *dev_id)
429{
430	struct net_device *dev = dev_id;
431	unsigned long e8390_base = dev->base_addr;
432	int interrupts, nr_serviced = 0;
433	struct ei_device *ei_local = netdev_priv(dev);
434
435	/*
436	 *	Protect the irq test too.
437	 */
438
439	spin_lock(&ei_local->page_lock);
440
441	if (ei_local->irqlock) {
442		/*
443		 * This might just be an interrupt for a PCI device sharing
444		 * this line
445		 */
446		netdev_err(dev, "Interrupted while interrupts are masked! isr=%#2x imr=%#2x\n",
447			   ei_inb_p(e8390_base + EN0_ISR),
448			   ei_inb_p(e8390_base + EN0_IMR));
449		spin_unlock(&ei_local->page_lock);
450		return IRQ_NONE;
451	}
452
453	/* Change to page 0 and read the intr status reg. */
454	ei_outb_p(E8390_NODMA+E8390_PAGE0, e8390_base + E8390_CMD);
455	netif_dbg(ei_local, intr, dev, "interrupt(isr=%#2.2x)\n",
456		  ei_inb_p(e8390_base + EN0_ISR));
457
458	/* !!Assumption!! -- we stay in page 0.	 Don't break this. */
459	while ((interrupts = ei_inb_p(e8390_base + EN0_ISR)) != 0 &&
460	       ++nr_serviced < MAX_SERVICE) {
461		if (!netif_running(dev)) {
462			netdev_warn(dev, "interrupt from stopped card\n");
463			/* rmk - acknowledge the interrupts */
464			ei_outb_p(interrupts, e8390_base + EN0_ISR);
465			interrupts = 0;
466			break;
467		}
468		if (interrupts & ENISR_OVER)
469			ei_rx_overrun(dev);
470		else if (interrupts & (ENISR_RX+ENISR_RX_ERR)) {
471			/* Got a good (?) packet. */
472			ei_receive(dev);
473		}
474		/* Push the next to-transmit packet through. */
475		if (interrupts & ENISR_TX)
476			ei_tx_intr(dev);
477		else if (interrupts & ENISR_TX_ERR)
478			ei_tx_err(dev);
479
480		if (interrupts & ENISR_COUNTERS) {
481			dev->stats.rx_frame_errors += ei_inb_p(e8390_base + EN0_COUNTER0);
482			dev->stats.rx_crc_errors   += ei_inb_p(e8390_base + EN0_COUNTER1);
483			dev->stats.rx_missed_errors += ei_inb_p(e8390_base + EN0_COUNTER2);
484			ei_outb_p(ENISR_COUNTERS, e8390_base + EN0_ISR); /* Ack intr. */
485		}
486
487		/* Ignore any RDC interrupts that make it back to here. */
488		if (interrupts & ENISR_RDC)
489			ei_outb_p(ENISR_RDC, e8390_base + EN0_ISR);
490
491		ei_outb_p(E8390_NODMA+E8390_PAGE0+E8390_START, e8390_base + E8390_CMD);
492	}
493
494	if (interrupts && (netif_msg_intr(ei_local))) {
495		ei_outb_p(E8390_NODMA+E8390_PAGE0+E8390_START, e8390_base + E8390_CMD);
496		if (nr_serviced >= MAX_SERVICE) {
497			/* 0xFF is valid for a card removal */
498			if (interrupts != 0xFF)
499				netdev_warn(dev, "Too much work at interrupt, status %#2.2x\n",
500					    interrupts);
501			ei_outb_p(ENISR_ALL, e8390_base + EN0_ISR); /* Ack. most intrs. */
502		} else {
503			netdev_warn(dev, "unknown interrupt %#2x\n", interrupts);
504			ei_outb_p(0xff, e8390_base + EN0_ISR); /* Ack. all intrs. */
505		}
506	}
507	spin_unlock(&ei_local->page_lock);
508	return IRQ_RETVAL(nr_serviced > 0);
509}
510
511#ifdef CONFIG_NET_POLL_CONTROLLER
512static void __ei_poll(struct net_device *dev)
513{
514	disable_irq(dev->irq);
515	__ei_interrupt(dev->irq, dev);
516	enable_irq(dev->irq);
517}
518#endif
519
520/**
521 * ei_tx_err - handle transmitter error
522 * @dev: network device which threw the exception
523 *
524 * A transmitter error has happened. Most likely excess collisions (which
525 * is a fairly normal condition). If the error is one where the Tx will
526 * have been aborted, we try and send another one right away, instead of
527 * letting the failed packet sit and collect dust in the Tx buffer. This
528 * is a much better solution as it avoids kernel based Tx timeouts, and
529 * an unnecessary card reset.
530 *
531 * Called with lock held.
532 */
533
534static void ei_tx_err(struct net_device *dev)
535{
536	unsigned long e8390_base = dev->base_addr;
537	/* ei_local is used on some platforms via the EI_SHIFT macro */
538	struct ei_device *ei_local __maybe_unused = netdev_priv(dev);
539	unsigned char txsr = ei_inb_p(e8390_base+EN0_TSR);
540	unsigned char tx_was_aborted = txsr & (ENTSR_ABT+ENTSR_FU);
541
542#ifdef VERBOSE_ERROR_DUMP
543	netdev_dbg(dev, "transmitter error (%#2x):", txsr);
544	if (txsr & ENTSR_ABT)
545		pr_cont(" excess-collisions ");
546	if (txsr & ENTSR_ND)
547		pr_cont(" non-deferral ");
548	if (txsr & ENTSR_CRS)
549		pr_cont(" lost-carrier ");
550	if (txsr & ENTSR_FU)
551		pr_cont(" FIFO-underrun ");
552	if (txsr & ENTSR_CDH)
553		pr_cont(" lost-heartbeat ");
554	pr_cont("\n");
555#endif
556
557	ei_outb_p(ENISR_TX_ERR, e8390_base + EN0_ISR); /* Ack intr. */
558
559	if (tx_was_aborted)
560		ei_tx_intr(dev);
561	else {
562		dev->stats.tx_errors++;
563		if (txsr & ENTSR_CRS)
564			dev->stats.tx_carrier_errors++;
565		if (txsr & ENTSR_CDH)
566			dev->stats.tx_heartbeat_errors++;
567		if (txsr & ENTSR_OWC)
568			dev->stats.tx_window_errors++;
569	}
570}
571
572/**
573 * ei_tx_intr - transmit interrupt handler
574 * @dev: network device for which tx intr is handled
575 *
576 * We have finished a transmit: check for errors and then trigger the next
577 * packet to be sent. Called with lock held.
578 */
579
580static void ei_tx_intr(struct net_device *dev)
581{
582	unsigned long e8390_base = dev->base_addr;
583	struct ei_device *ei_local = netdev_priv(dev);
584	int status = ei_inb(e8390_base + EN0_TSR);
585
586	ei_outb_p(ENISR_TX, e8390_base + EN0_ISR); /* Ack intr. */
587
588	/*
589	 * There are two Tx buffers, see which one finished, and trigger
590	 * the send of another one if it exists.
591	 */
592	ei_local->txqueue--;
593
594	if (ei_local->tx1 < 0) {
595		if (ei_local->lasttx != 1 && ei_local->lasttx != -1)
596			pr_err("%s: bogus last_tx_buffer %d, tx1=%d\n",
597			       ei_local->name, ei_local->lasttx, ei_local->tx1);
598		ei_local->tx1 = 0;
599		if (ei_local->tx2 > 0) {
600			ei_local->txing = 1;
601			NS8390_trigger_send(dev, ei_local->tx2, ei_local->tx_start_page + 6);
602			netif_trans_update(dev);
603			ei_local->tx2 = -1;
604			ei_local->lasttx = 2;
605		} else {
606			ei_local->lasttx = 20;
607			ei_local->txing = 0;
608		}
609	} else if (ei_local->tx2 < 0) {
610		if (ei_local->lasttx != 2  &&  ei_local->lasttx != -2)
611			pr_err("%s: bogus last_tx_buffer %d, tx2=%d\n",
612			       ei_local->name, ei_local->lasttx, ei_local->tx2);
613		ei_local->tx2 = 0;
614		if (ei_local->tx1 > 0) {
615			ei_local->txing = 1;
616			NS8390_trigger_send(dev, ei_local->tx1, ei_local->tx_start_page);
617			netif_trans_update(dev);
618			ei_local->tx1 = -1;
619			ei_local->lasttx = 1;
620		} else {
621			ei_local->lasttx = 10;
622			ei_local->txing = 0;
623		}
624	} /* else
625		netdev_warn(dev, "unexpected TX-done interrupt, lasttx=%d\n",
626			    ei_local->lasttx);
627*/
628
629	/* Minimize Tx latency: update the statistics after we restart TXing. */
630	if (status & ENTSR_COL)
631		dev->stats.collisions++;
632	if (status & ENTSR_PTX)
633		dev->stats.tx_packets++;
634	else {
635		dev->stats.tx_errors++;
636		if (status & ENTSR_ABT) {
637			dev->stats.tx_aborted_errors++;
638			dev->stats.collisions += 16;
639		}
640		if (status & ENTSR_CRS)
641			dev->stats.tx_carrier_errors++;
642		if (status & ENTSR_FU)
643			dev->stats.tx_fifo_errors++;
644		if (status & ENTSR_CDH)
645			dev->stats.tx_heartbeat_errors++;
646		if (status & ENTSR_OWC)
647			dev->stats.tx_window_errors++;
648	}
649	netif_wake_queue(dev);
650}
651
652/**
653 * ei_receive - receive some packets
654 * @dev: network device with which receive will be run
655 *
656 * We have a good packet(s), get it/them out of the buffers.
657 * Called with lock held.
658 */
659
660static void ei_receive(struct net_device *dev)
661{
662	unsigned long e8390_base = dev->base_addr;
663	struct ei_device *ei_local = netdev_priv(dev);
664	unsigned char rxing_page, this_frame, next_frame;
665	unsigned short current_offset;
666	int rx_pkt_count = 0;
667	struct e8390_pkt_hdr rx_frame;
668	int num_rx_pages = ei_local->stop_page-ei_local->rx_start_page;
669
670	while (++rx_pkt_count < 10) {
671		int pkt_len, pkt_stat;
672
673		/* Get the rx page (incoming packet pointer). */
674		ei_outb_p(E8390_NODMA+E8390_PAGE1, e8390_base + E8390_CMD);
675		rxing_page = ei_inb_p(e8390_base + EN1_CURPAG);
676		ei_outb_p(E8390_NODMA+E8390_PAGE0, e8390_base + E8390_CMD);
677
678		/* Remove one frame from the ring.  Boundary is always a page behind. */
679		this_frame = ei_inb_p(e8390_base + EN0_BOUNDARY) + 1;
680		if (this_frame >= ei_local->stop_page)
681			this_frame = ei_local->rx_start_page;
682
683		/* Someday we'll omit the previous, iff we never get this message.
684		   (There is at least one clone claimed to have a problem.)
685
686		   Keep quiet if it looks like a card removal. One problem here
687		   is that some clones crash in roughly the same way.
688		 */
689		if ((netif_msg_rx_status(ei_local)) &&
690		    this_frame != ei_local->current_page &&
691		    (this_frame != 0x0 || rxing_page != 0xFF))
692			netdev_err(dev,
693				   "mismatched read page pointers %2x vs %2x\n",
694				   this_frame, ei_local->current_page);
695
696		if (this_frame == rxing_page)	/* Read all the frames? */
697			break;				/* Done for now */
698
699		current_offset = this_frame << 8;
700		ei_get_8390_hdr(dev, &rx_frame, this_frame);
701
702		pkt_len = rx_frame.count - sizeof(struct e8390_pkt_hdr);
703		pkt_stat = rx_frame.status;
704
705		next_frame = this_frame + 1 + ((pkt_len+4)>>8);
706
707		/* Check for bogosity warned by 3c503 book: the status byte is never
708		   written.  This happened a lot during testing! This code should be
709		   cleaned up someday. */
710		if (rx_frame.next != next_frame &&
711		    rx_frame.next != next_frame + 1 &&
712		    rx_frame.next != next_frame - num_rx_pages &&
713		    rx_frame.next != next_frame + 1 - num_rx_pages) {
714			ei_local->current_page = rxing_page;
715			ei_outb(ei_local->current_page-1, e8390_base+EN0_BOUNDARY);
716			dev->stats.rx_errors++;
717			continue;
718		}
719
720		if (pkt_len < 60  ||  pkt_len > 1518) {
721			netif_dbg(ei_local, rx_status, dev,
722				  "bogus packet size: %d, status=%#2x nxpg=%#2x\n",
723				  rx_frame.count, rx_frame.status,
724				  rx_frame.next);
725			dev->stats.rx_errors++;
726			dev->stats.rx_length_errors++;
727		} else if ((pkt_stat & 0x0F) == ENRSR_RXOK) {
728			struct sk_buff *skb;
729
730			skb = netdev_alloc_skb(dev, pkt_len + 2);
731			if (skb == NULL) {
732				netif_err(ei_local, rx_err, dev,
733					  "Couldn't allocate a sk_buff of size %d\n",
734					  pkt_len);
735				dev->stats.rx_dropped++;
736				break;
737			} else {
738				skb_reserve(skb, 2);	/* IP headers on 16 byte boundaries */
739				skb_put(skb, pkt_len);	/* Make room */
740				ei_block_input(dev, pkt_len, skb, current_offset + sizeof(rx_frame));
741				skb->protocol = eth_type_trans(skb, dev);
742				if (!skb_defer_rx_timestamp(skb))
743					netif_rx(skb);
744				dev->stats.rx_packets++;
745				dev->stats.rx_bytes += pkt_len;
746				if (pkt_stat & ENRSR_PHY)
747					dev->stats.multicast++;
748			}
749		} else {
750			netif_err(ei_local, rx_err, dev,
751				  "bogus packet: status=%#2x nxpg=%#2x size=%d\n",
752				  rx_frame.status, rx_frame.next,
753				  rx_frame.count);
754			dev->stats.rx_errors++;
755			/* NB: The NIC counts CRC, frame and missed errors. */
756			if (pkt_stat & ENRSR_FO)
757				dev->stats.rx_fifo_errors++;
758		}
759		next_frame = rx_frame.next;
760
761		/* This _should_ never happen: it's here for avoiding bad clones. */
762		if (next_frame >= ei_local->stop_page) {
763			netdev_notice(dev, "next frame inconsistency, %#2x\n",
764				      next_frame);
765			next_frame = ei_local->rx_start_page;
766		}
767		ei_local->current_page = next_frame;
768		ei_outb_p(next_frame-1, e8390_base+EN0_BOUNDARY);
769	}
770
771	/* We used to also ack ENISR_OVER here, but that would sometimes mask
772	   a real overrun, leaving the 8390 in a stopped state with rec'vr off. */
773	ei_outb_p(ENISR_RX+ENISR_RX_ERR, e8390_base+EN0_ISR);
774}
775
776/**
777 * ei_rx_overrun - handle receiver overrun
778 * @dev: network device which threw exception
779 *
780 * We have a receiver overrun: we have to kick the 8390 to get it started
781 * again. Problem is that you have to kick it exactly as NS prescribes in
782 * the updated datasheets, or "the NIC may act in an unpredictable manner."
783 * This includes causing "the NIC to defer indefinitely when it is stopped
784 * on a busy network."  Ugh.
785 * Called with lock held. Don't call this with the interrupts off or your
786 * computer will hate you - it takes 10ms or so.
787 */
788
789static void ei_rx_overrun(struct net_device *dev)
790{
791	unsigned long e8390_base = dev->base_addr;
792	unsigned char was_txing, must_resend = 0;
793	/* ei_local is used on some platforms via the EI_SHIFT macro */
794	struct ei_device *ei_local __maybe_unused = netdev_priv(dev);
795
796	/*
797	 * Record whether a Tx was in progress and then issue the
798	 * stop command.
799	 */
800	was_txing = ei_inb_p(e8390_base+E8390_CMD) & E8390_TRANS;
801	ei_outb_p(E8390_NODMA+E8390_PAGE0+E8390_STOP, e8390_base+E8390_CMD);
802
803	netif_dbg(ei_local, rx_err, dev, "Receiver overrun\n");
804	dev->stats.rx_over_errors++;
805
806	/*
807	 * Wait a full Tx time (1.2ms) + some guard time, NS says 1.6ms total.
808	 * Early datasheets said to poll the reset bit, but now they say that
809	 * it "is not a reliable indicator and subsequently should be ignored."
810	 * We wait at least 10ms.
811	 */
812
813	mdelay(10);
814
815	/*
816	 * Reset RBCR[01] back to zero as per magic incantation.
817	 */
818	ei_outb_p(0x00, e8390_base+EN0_RCNTLO);
819	ei_outb_p(0x00, e8390_base+EN0_RCNTHI);
820
821	/*
822	 * See if any Tx was interrupted or not. According to NS, this
823	 * step is vital, and skipping it will cause no end of havoc.
824	 */
825
826	if (was_txing) {
827		unsigned char tx_completed = ei_inb_p(e8390_base+EN0_ISR) & (ENISR_TX+ENISR_TX_ERR);
828		if (!tx_completed)
829			must_resend = 1;
830	}
831
832	/*
833	 * Have to enter loopback mode and then restart the NIC before
834	 * you are allowed to slurp packets up off the ring.
835	 */
836	ei_outb_p(E8390_TXOFF, e8390_base + EN0_TXCR);
837	ei_outb_p(E8390_NODMA + E8390_PAGE0 + E8390_START, e8390_base + E8390_CMD);
838
839	/*
840	 * Clear the Rx ring of all the debris, and ack the interrupt.
841	 */
842	ei_receive(dev);
843	ei_outb_p(ENISR_OVER, e8390_base+EN0_ISR);
844
845	/*
846	 * Leave loopback mode, and resend any packet that got stopped.
847	 */
848	ei_outb_p(E8390_TXCONFIG, e8390_base + EN0_TXCR);
849	if (must_resend)
850		ei_outb_p(E8390_NODMA + E8390_PAGE0 + E8390_START + E8390_TRANS, e8390_base + E8390_CMD);
851}
852
853/*
854 *	Collect the stats. This is called unlocked and from several contexts.
855 */
856
857static struct net_device_stats *__ei_get_stats(struct net_device *dev)
858{
859	unsigned long ioaddr = dev->base_addr;
860	struct ei_device *ei_local = netdev_priv(dev);
861	unsigned long flags;
862
863	/* If the card is stopped, just return the present stats. */
864	if (!netif_running(dev))
865		return &dev->stats;
866
867	spin_lock_irqsave(&ei_local->page_lock, flags);
868	/* Read the counter registers, assuming we are in page 0. */
869	dev->stats.rx_frame_errors  += ei_inb_p(ioaddr + EN0_COUNTER0);
870	dev->stats.rx_crc_errors    += ei_inb_p(ioaddr + EN0_COUNTER1);
871	dev->stats.rx_missed_errors += ei_inb_p(ioaddr + EN0_COUNTER2);
872	spin_unlock_irqrestore(&ei_local->page_lock, flags);
873
874	return &dev->stats;
875}
876
877/*
878 * Form the 64 bit 8390 multicast table from the linked list of addresses
879 * associated with this dev structure.
880 */
881
882static inline void make_mc_bits(u8 *bits, struct net_device *dev)
883{
884	struct netdev_hw_addr *ha;
885
886	netdev_for_each_mc_addr(ha, dev) {
887		u32 crc = ether_crc(ETH_ALEN, ha->addr);
888		/*
889		 * The 8390 uses the 6 most significant bits of the
890		 * CRC to index the multicast table.
891		 */
892		bits[crc>>29] |= (1<<((crc>>26)&7));
893	}
894}
895
896/**
897 * do_set_multicast_list - set/clear multicast filter
898 * @dev: net device for which multicast filter is adjusted
899 *
900 *	Set or clear the multicast filter for this adaptor. May be called
901 *	from a BH in 2.1.x. Must be called with lock held.
902 */
903
904static void do_set_multicast_list(struct net_device *dev)
905{
906	unsigned long e8390_base = dev->base_addr;
907	int i;
908	struct ei_device *ei_local = netdev_priv(dev);
909
910	if (!(dev->flags&(IFF_PROMISC|IFF_ALLMULTI))) {
911		memset(ei_local->mcfilter, 0, 8);
912		if (!netdev_mc_empty(dev))
913			make_mc_bits(ei_local->mcfilter, dev);
914	} else
915		memset(ei_local->mcfilter, 0xFF, 8);	/* mcast set to accept-all */
916
917	/*
918	 * DP8390 manuals don't specify any magic sequence for altering
919	 * the multicast regs on an already running card. To be safe, we
920	 * ensure multicast mode is off prior to loading up the new hash
921	 * table. If this proves to be not enough, we can always resort
922	 * to stopping the NIC, loading the table and then restarting.
923	 *
924	 * Bug Alert!  The MC regs on the SMC 83C690 (SMC Elite and SMC
925	 * Elite16) appear to be write-only. The NS 8390 data sheet lists
926	 * them as r/w so this is a bug.  The SMC 83C790 (SMC Ultra and
927	 * Ultra32 EISA) appears to have this bug fixed.
928	 */
929
930	if (netif_running(dev))
931		ei_outb_p(E8390_RXCONFIG, e8390_base + EN0_RXCR);
932	ei_outb_p(E8390_NODMA + E8390_PAGE1, e8390_base + E8390_CMD);
933	for (i = 0; i < 8; i++) {
934		ei_outb_p(ei_local->mcfilter[i], e8390_base + EN1_MULT_SHIFT(i));
935#ifndef BUG_83C690
936		if (ei_inb_p(e8390_base + EN1_MULT_SHIFT(i)) != ei_local->mcfilter[i])
937			netdev_err(dev, "Multicast filter read/write mismap %d\n",
938				   i);
939#endif
940	}
941	ei_outb_p(E8390_NODMA + E8390_PAGE0, e8390_base + E8390_CMD);
942
943	if (dev->flags&IFF_PROMISC)
944		ei_outb_p(E8390_RXCONFIG | 0x18, e8390_base + EN0_RXCR);
945	else if (dev->flags & IFF_ALLMULTI || !netdev_mc_empty(dev))
946		ei_outb_p(E8390_RXCONFIG | 0x08, e8390_base + EN0_RXCR);
947	else
948		ei_outb_p(E8390_RXCONFIG, e8390_base + EN0_RXCR);
949}
950
951/*
952 *	Called without lock held. This is invoked from user context and may
953 *	be parallel to just about everything else. Its also fairly quick and
954 *	not called too often. Must protect against both bh and irq users
955 */
956
957static void __ei_set_multicast_list(struct net_device *dev)
958{
959	unsigned long flags;
960	struct ei_device *ei_local = netdev_priv(dev);
961
962	spin_lock_irqsave(&ei_local->page_lock, flags);
963	do_set_multicast_list(dev);
964	spin_unlock_irqrestore(&ei_local->page_lock, flags);
965}
966
967/**
968 * ethdev_setup - init rest of 8390 device struct
969 * @dev: network device structure to init
970 *
971 * Initialize the rest of the 8390 device structure.  Do NOT __init
972 * this, as it is used by 8390 based modular drivers too.
973 */
974
975static void ethdev_setup(struct net_device *dev)
976{
977	struct ei_device *ei_local = netdev_priv(dev);
978
979	ether_setup(dev);
980
981	spin_lock_init(&ei_local->page_lock);
982
983	ei_local->msg_enable = netif_msg_init(msg_enable, default_msg_level);
984
985	if (netif_msg_drv(ei_local) && (version_printed++ == 0))
986		pr_info("%s", version);
987}
988
989/**
990 * alloc_ei_netdev - alloc_etherdev counterpart for 8390
991 * @size: extra bytes to allocate
992 *
993 * Allocate 8390-specific net_device.
994 */
995static struct net_device *____alloc_ei_netdev(int size)
996{
997	return alloc_netdev(sizeof(struct ei_device) + size, "eth%d",
998			    NET_NAME_UNKNOWN, ethdev_setup);
999}
1000
1001
1002
1003
1004/* This page of functions should be 8390 generic */
1005/* Follow National Semi's recommendations for initializing the "NIC". */
1006
1007/**
1008 * NS8390_init - initialize 8390 hardware
1009 * @dev: network device to initialize
1010 * @startp: boolean.  non-zero value to initiate chip processing
1011 *
1012 *	Must be called with lock held.
1013 */
1014
1015static void __NS8390_init(struct net_device *dev, int startp)
1016{
1017	unsigned long e8390_base = dev->base_addr;
1018	struct ei_device *ei_local = netdev_priv(dev);
1019	int i;
1020	int endcfg = ei_local->word16
1021	    ? (0x48 | ENDCFG_WTS | (ei_local->bigendian ? ENDCFG_BOS : 0))
1022	    : 0x48;
1023
1024	BUILD_BUG_ON(sizeof(struct e8390_pkt_hdr) != 4);
1025	/* Follow National Semi's recommendations for initing the DP83902. */
1026	ei_outb_p(E8390_NODMA+E8390_PAGE0+E8390_STOP, e8390_base+E8390_CMD); /* 0x21 */
1027	ei_outb_p(endcfg, e8390_base + EN0_DCFG);	/* 0x48 or 0x49 */
1028	/* Clear the remote byte count registers. */
1029	ei_outb_p(0x00,  e8390_base + EN0_RCNTLO);
1030	ei_outb_p(0x00,  e8390_base + EN0_RCNTHI);
1031	/* Set to monitor and loopback mode -- this is vital!. */
1032	ei_outb_p(E8390_RXOFF, e8390_base + EN0_RXCR); /* 0x20 */
1033	ei_outb_p(E8390_TXOFF, e8390_base + EN0_TXCR); /* 0x02 */
1034	/* Set the transmit page and receive ring. */
1035	ei_outb_p(ei_local->tx_start_page, e8390_base + EN0_TPSR);
1036	ei_local->tx1 = ei_local->tx2 = 0;
1037	ei_outb_p(ei_local->rx_start_page, e8390_base + EN0_STARTPG);
1038	ei_outb_p(ei_local->stop_page-1, e8390_base + EN0_BOUNDARY);	/* 3c503 says 0x3f,NS0x26*/
1039	ei_local->current_page = ei_local->rx_start_page;		/* assert boundary+1 */
1040	ei_outb_p(ei_local->stop_page, e8390_base + EN0_STOPPG);
1041	/* Clear the pending interrupts and mask. */
1042	ei_outb_p(0xFF, e8390_base + EN0_ISR);
1043	ei_outb_p(0x00,  e8390_base + EN0_IMR);
1044
1045	/* Copy the station address into the DS8390 registers. */
1046
1047	ei_outb_p(E8390_NODMA + E8390_PAGE1 + E8390_STOP, e8390_base+E8390_CMD); /* 0x61 */
1048	for (i = 0; i < 6; i++) {
1049		ei_outb_p(dev->dev_addr[i], e8390_base + EN1_PHYS_SHIFT(i));
1050		if ((netif_msg_probe(ei_local)) &&
1051		    ei_inb_p(e8390_base + EN1_PHYS_SHIFT(i)) != dev->dev_addr[i])
1052			netdev_err(dev,
1053				   "Hw. address read/write mismap %d\n", i);
1054	}
1055
1056	ei_outb_p(ei_local->rx_start_page, e8390_base + EN1_CURPAG);
1057	ei_outb_p(E8390_NODMA+E8390_PAGE0+E8390_STOP, e8390_base+E8390_CMD);
1058
1059	ei_local->tx1 = ei_local->tx2 = 0;
1060	ei_local->txing = 0;
1061
1062	if (startp) {
1063		ei_outb_p(0xff,  e8390_base + EN0_ISR);
1064		ei_outb_p(ENISR_ALL,  e8390_base + EN0_IMR);
1065		ei_outb_p(E8390_NODMA+E8390_PAGE0+E8390_START, e8390_base+E8390_CMD);
1066		ei_outb_p(E8390_TXCONFIG, e8390_base + EN0_TXCR); /* xmit on. */
1067		/* 3c503 TechMan says rxconfig only after the NIC is started. */
1068		ei_outb_p(E8390_RXCONFIG, e8390_base + EN0_RXCR); /* rx on,  */
1069		do_set_multicast_list(dev);	/* (re)load the mcast table */
1070	}
1071}
1072
1073/* Trigger a transmit start, assuming the length is valid.
1074   Always called with the page lock held */
1075
1076static void NS8390_trigger_send(struct net_device *dev, unsigned int length,
1077								int start_page)
1078{
1079	unsigned long e8390_base = dev->base_addr;
1080	struct ei_device *ei_local __attribute((unused)) = netdev_priv(dev);
1081
1082	ei_outb_p(E8390_NODMA+E8390_PAGE0, e8390_base+E8390_CMD);
1083
1084	if (ei_inb_p(e8390_base + E8390_CMD) & E8390_TRANS) {
1085		netdev_warn(dev, "trigger_send() called with the transmitter busy\n");
1086		return;
1087	}
1088	ei_outb_p(length & 0xff, e8390_base + EN0_TCNTLO);
1089	ei_outb_p(length >> 8, e8390_base + EN0_TCNTHI);
1090	ei_outb_p(start_page, e8390_base + EN0_TPSR);
1091	ei_outb_p(E8390_NODMA+E8390_TRANS+E8390_START, e8390_base+E8390_CMD);
1092}
1093