Tulip throws an error when dma debugging is enabled, as it doesn't properly check dma mapping results with dma_mapping_error() durring tx ring refills. Easy fix, just add it in, and drop the frame if the mapping is bad Signed-off-by: Neil Horman <nhorman@tuxdriver.com> CC: Grant Grundler <grundler@parisc-linux.org> CC: "David S. Miller" <davem@davemloft.net> Reviewed-by: Grant Grundler <grundler@parisc-linux.org> Signed-off-by: David S. Miller <davem@davemloft.net>
		
			
				
	
	
		
			814 lines
		
	
	
	
		
			25 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			814 lines
		
	
	
	
		
			25 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
	drivers/net/ethernet/dec/tulip/interrupt.c
 | 
						|
 | 
						|
	Copyright 2000,2001  The Linux Kernel Team
 | 
						|
	Written/copyright 1994-2001 by Donald Becker.
 | 
						|
 | 
						|
	This software may be used and distributed according to the terms
 | 
						|
	of the GNU General Public License, incorporated herein by reference.
 | 
						|
 | 
						|
        Please submit bugs to http://bugzilla.kernel.org/ .
 | 
						|
*/
 | 
						|
 | 
						|
#include <linux/pci.h>
 | 
						|
#include "tulip.h"
 | 
						|
#include <linux/etherdevice.h>
 | 
						|
 | 
						|
int tulip_rx_copybreak;
 | 
						|
unsigned int tulip_max_interrupt_work;
 | 
						|
 | 
						|
#ifdef CONFIG_TULIP_NAPI_HW_MITIGATION
 | 
						|
#define MIT_SIZE 15
 | 
						|
#define MIT_TABLE 15 /* We use 0 or max */
 | 
						|
 | 
						|
static unsigned int mit_table[MIT_SIZE+1] =
 | 
						|
{
 | 
						|
        /*  CRS11 21143 hardware Mitigation Control Interrupt
 | 
						|
            We use only RX mitigation we other techniques for
 | 
						|
            TX intr. mitigation.
 | 
						|
 | 
						|
           31    Cycle Size (timer control)
 | 
						|
           30:27 TX timer in 16 * Cycle size
 | 
						|
           26:24 TX No pkts before Int.
 | 
						|
           23:20 RX timer in Cycle size
 | 
						|
           19:17 RX No pkts before Int.
 | 
						|
           16       Continues Mode (CM)
 | 
						|
        */
 | 
						|
 | 
						|
        0x0,             /* IM disabled */
 | 
						|
        0x80150000,      /* RX time = 1, RX pkts = 2, CM = 1 */
 | 
						|
        0x80150000,
 | 
						|
        0x80270000,
 | 
						|
        0x80370000,
 | 
						|
        0x80490000,
 | 
						|
        0x80590000,
 | 
						|
        0x80690000,
 | 
						|
        0x807B0000,
 | 
						|
        0x808B0000,
 | 
						|
        0x809D0000,
 | 
						|
        0x80AD0000,
 | 
						|
        0x80BD0000,
 | 
						|
        0x80CF0000,
 | 
						|
        0x80DF0000,
 | 
						|
//       0x80FF0000      /* RX time = 16, RX pkts = 7, CM = 1 */
 | 
						|
        0x80F10000      /* RX time = 16, RX pkts = 0, CM = 1 */
 | 
						|
};
 | 
						|
#endif
 | 
						|
 | 
						|
 | 
						|
int tulip_refill_rx(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct tulip_private *tp = netdev_priv(dev);
 | 
						|
	int entry;
 | 
						|
	int refilled = 0;
 | 
						|
 | 
						|
	/* Refill the Rx ring buffers. */
 | 
						|
	for (; tp->cur_rx - tp->dirty_rx > 0; tp->dirty_rx++) {
 | 
						|
		entry = tp->dirty_rx % RX_RING_SIZE;
 | 
						|
		if (tp->rx_buffers[entry].skb == NULL) {
 | 
						|
			struct sk_buff *skb;
 | 
						|
			dma_addr_t mapping;
 | 
						|
 | 
						|
			skb = tp->rx_buffers[entry].skb =
 | 
						|
				netdev_alloc_skb(dev, PKT_BUF_SZ);
 | 
						|
			if (skb == NULL)
 | 
						|
				break;
 | 
						|
 | 
						|
			mapping = pci_map_single(tp->pdev, skb->data, PKT_BUF_SZ,
 | 
						|
						 PCI_DMA_FROMDEVICE);
 | 
						|
			if (dma_mapping_error(&tp->pdev->dev, mapping)) {
 | 
						|
				dev_kfree_skb(skb);
 | 
						|
				tp->rx_buffers[entry].skb = NULL;
 | 
						|
				break;
 | 
						|
			}
 | 
						|
 | 
						|
			tp->rx_buffers[entry].mapping = mapping;
 | 
						|
 | 
						|
			tp->rx_ring[entry].buffer1 = cpu_to_le32(mapping);
 | 
						|
			refilled++;
 | 
						|
		}
 | 
						|
		tp->rx_ring[entry].status = cpu_to_le32(DescOwned);
 | 
						|
	}
 | 
						|
	if(tp->chip_id == LC82C168) {
 | 
						|
		if(((ioread32(tp->base_addr + CSR5)>>17)&0x07) == 4) {
 | 
						|
			/* Rx stopped due to out of buffers,
 | 
						|
			 * restart it
 | 
						|
			 */
 | 
						|
			iowrite32(0x01, tp->base_addr + CSR2);
 | 
						|
		}
 | 
						|
	}
 | 
						|
	return refilled;
 | 
						|
}
 | 
						|
 | 
						|
#ifdef CONFIG_TULIP_NAPI
 | 
						|
 | 
						|
void oom_timer(unsigned long data)
 | 
						|
{
 | 
						|
        struct net_device *dev = (struct net_device *)data;
 | 
						|
	struct tulip_private *tp = netdev_priv(dev);
 | 
						|
	napi_schedule(&tp->napi);
 | 
						|
}
 | 
						|
 | 
						|
int tulip_poll(struct napi_struct *napi, int budget)
 | 
						|
{
 | 
						|
	struct tulip_private *tp = container_of(napi, struct tulip_private, napi);
 | 
						|
	struct net_device *dev = tp->dev;
 | 
						|
	int entry = tp->cur_rx % RX_RING_SIZE;
 | 
						|
	int work_done = 0;
 | 
						|
#ifdef CONFIG_TULIP_NAPI_HW_MITIGATION
 | 
						|
	int received = 0;
 | 
						|
#endif
 | 
						|
 | 
						|
#ifdef CONFIG_TULIP_NAPI_HW_MITIGATION
 | 
						|
 | 
						|
/* that one buffer is needed for mit activation; or might be a
 | 
						|
   bug in the ring buffer code; check later -- JHS*/
 | 
						|
 | 
						|
        if (budget >=RX_RING_SIZE) budget--;
 | 
						|
#endif
 | 
						|
 | 
						|
	if (tulip_debug > 4)
 | 
						|
		netdev_dbg(dev, " In tulip_rx(), entry %d %08x\n",
 | 
						|
			   entry, tp->rx_ring[entry].status);
 | 
						|
 | 
						|
       do {
 | 
						|
		if (ioread32(tp->base_addr + CSR5) == 0xffffffff) {
 | 
						|
			netdev_dbg(dev, " In tulip_poll(), hardware disappeared\n");
 | 
						|
			break;
 | 
						|
		}
 | 
						|
               /* Acknowledge current RX interrupt sources. */
 | 
						|
               iowrite32((RxIntr | RxNoBuf), tp->base_addr + CSR5);
 | 
						|
 | 
						|
 | 
						|
               /* If we own the next entry, it is a new packet. Send it up. */
 | 
						|
               while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) {
 | 
						|
                       s32 status = le32_to_cpu(tp->rx_ring[entry].status);
 | 
						|
		       short pkt_len;
 | 
						|
 | 
						|
                       if (tp->dirty_rx + RX_RING_SIZE == tp->cur_rx)
 | 
						|
                               break;
 | 
						|
 | 
						|
		       if (tulip_debug > 5)
 | 
						|
				netdev_dbg(dev, "In tulip_rx(), entry %d %08x\n",
 | 
						|
					   entry, status);
 | 
						|
 | 
						|
		       if (++work_done >= budget)
 | 
						|
                               goto not_done;
 | 
						|
 | 
						|
		       /*
 | 
						|
			* Omit the four octet CRC from the length.
 | 
						|
			* (May not be considered valid until we have
 | 
						|
			* checked status for RxLengthOver2047 bits)
 | 
						|
			*/
 | 
						|
		       pkt_len = ((status >> 16) & 0x7ff) - 4;
 | 
						|
 | 
						|
		       /*
 | 
						|
			* Maximum pkt_len is 1518 (1514 + vlan header)
 | 
						|
			* Anything higher than this is always invalid
 | 
						|
			* regardless of RxLengthOver2047 bits
 | 
						|
			*/
 | 
						|
 | 
						|
		       if ((status & (RxLengthOver2047 |
 | 
						|
				      RxDescCRCError |
 | 
						|
				      RxDescCollisionSeen |
 | 
						|
				      RxDescRunt |
 | 
						|
				      RxDescDescErr |
 | 
						|
				      RxWholePkt)) != RxWholePkt ||
 | 
						|
			   pkt_len > 1518) {
 | 
						|
			       if ((status & (RxLengthOver2047 |
 | 
						|
					      RxWholePkt)) != RxWholePkt) {
 | 
						|
                                /* Ingore earlier buffers. */
 | 
						|
                                       if ((status & 0xffff) != 0x7fff) {
 | 
						|
                                               if (tulip_debug > 1)
 | 
						|
                                                       dev_warn(&dev->dev,
 | 
						|
								"Oversized Ethernet frame spanned multiple buffers, status %08x!\n",
 | 
						|
								status);
 | 
						|
						dev->stats.rx_length_errors++;
 | 
						|
					}
 | 
						|
			       } else {
 | 
						|
                                /* There was a fatal error. */
 | 
						|
				       if (tulip_debug > 2)
 | 
						|
						netdev_dbg(dev, "Receive error, Rx status %08x\n",
 | 
						|
							   status);
 | 
						|
					dev->stats.rx_errors++; /* end of a packet.*/
 | 
						|
					if (pkt_len > 1518 ||
 | 
						|
					    (status & RxDescRunt))
 | 
						|
						dev->stats.rx_length_errors++;
 | 
						|
 | 
						|
					if (status & 0x0004)
 | 
						|
						dev->stats.rx_frame_errors++;
 | 
						|
					if (status & 0x0002)
 | 
						|
						dev->stats.rx_crc_errors++;
 | 
						|
					if (status & 0x0001)
 | 
						|
						dev->stats.rx_fifo_errors++;
 | 
						|
                               }
 | 
						|
                       } else {
 | 
						|
                               struct sk_buff *skb;
 | 
						|
 | 
						|
                               /* Check if the packet is long enough to accept without copying
 | 
						|
                                  to a minimally-sized skbuff. */
 | 
						|
                               if (pkt_len < tulip_rx_copybreak &&
 | 
						|
                                   (skb = netdev_alloc_skb(dev, pkt_len + 2)) != NULL) {
 | 
						|
                                       skb_reserve(skb, 2);    /* 16 byte align the IP header */
 | 
						|
                                       pci_dma_sync_single_for_cpu(tp->pdev,
 | 
						|
								   tp->rx_buffers[entry].mapping,
 | 
						|
								   pkt_len, PCI_DMA_FROMDEVICE);
 | 
						|
#if ! defined(__alpha__)
 | 
						|
                                       skb_copy_to_linear_data(skb, tp->rx_buffers[entry].skb->data,
 | 
						|
                                                        pkt_len);
 | 
						|
                                       skb_put(skb, pkt_len);
 | 
						|
#else
 | 
						|
                                       memcpy(skb_put(skb, pkt_len),
 | 
						|
                                              tp->rx_buffers[entry].skb->data,
 | 
						|
                                              pkt_len);
 | 
						|
#endif
 | 
						|
                                       pci_dma_sync_single_for_device(tp->pdev,
 | 
						|
								      tp->rx_buffers[entry].mapping,
 | 
						|
								      pkt_len, PCI_DMA_FROMDEVICE);
 | 
						|
                               } else {        /* Pass up the skb already on the Rx ring. */
 | 
						|
                                       char *temp = skb_put(skb = tp->rx_buffers[entry].skb,
 | 
						|
                                                            pkt_len);
 | 
						|
 | 
						|
#ifndef final_version
 | 
						|
                                       if (tp->rx_buffers[entry].mapping !=
 | 
						|
                                           le32_to_cpu(tp->rx_ring[entry].buffer1)) {
 | 
						|
                                               dev_err(&dev->dev,
 | 
						|
						       "Internal fault: The skbuff addresses do not match in tulip_rx: %08x vs. %08llx %p / %p\n",
 | 
						|
						       le32_to_cpu(tp->rx_ring[entry].buffer1),
 | 
						|
						       (unsigned long long)tp->rx_buffers[entry].mapping,
 | 
						|
						       skb->head, temp);
 | 
						|
                                       }
 | 
						|
#endif
 | 
						|
 | 
						|
                                       pci_unmap_single(tp->pdev, tp->rx_buffers[entry].mapping,
 | 
						|
                                                        PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
 | 
						|
 | 
						|
                                       tp->rx_buffers[entry].skb = NULL;
 | 
						|
                                       tp->rx_buffers[entry].mapping = 0;
 | 
						|
                               }
 | 
						|
                               skb->protocol = eth_type_trans(skb, dev);
 | 
						|
 | 
						|
                               netif_receive_skb(skb);
 | 
						|
 | 
						|
				dev->stats.rx_packets++;
 | 
						|
				dev->stats.rx_bytes += pkt_len;
 | 
						|
                       }
 | 
						|
#ifdef CONFIG_TULIP_NAPI_HW_MITIGATION
 | 
						|
		       received++;
 | 
						|
#endif
 | 
						|
 | 
						|
                       entry = (++tp->cur_rx) % RX_RING_SIZE;
 | 
						|
                       if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/4)
 | 
						|
                               tulip_refill_rx(dev);
 | 
						|
 | 
						|
                }
 | 
						|
 | 
						|
               /* New ack strategy... irq does not ack Rx any longer
 | 
						|
                  hopefully this helps */
 | 
						|
 | 
						|
               /* Really bad things can happen here... If new packet arrives
 | 
						|
                * and an irq arrives (tx or just due to occasionally unset
 | 
						|
                * mask), it will be acked by irq handler, but new thread
 | 
						|
                * is not scheduled. It is major hole in design.
 | 
						|
                * No idea how to fix this if "playing with fire" will fail
 | 
						|
                * tomorrow (night 011029). If it will not fail, we won
 | 
						|
                * finally: amount of IO did not increase at all. */
 | 
						|
       } while ((ioread32(tp->base_addr + CSR5) & RxIntr));
 | 
						|
 | 
						|
 #ifdef CONFIG_TULIP_NAPI_HW_MITIGATION
 | 
						|
 | 
						|
          /* We use this simplistic scheme for IM. It's proven by
 | 
						|
             real life installations. We can have IM enabled
 | 
						|
            continuesly but this would cause unnecessary latency.
 | 
						|
            Unfortunely we can't use all the NET_RX_* feedback here.
 | 
						|
            This would turn on IM for devices that is not contributing
 | 
						|
            to backlog congestion with unnecessary latency.
 | 
						|
 | 
						|
             We monitor the device RX-ring and have:
 | 
						|
 | 
						|
             HW Interrupt Mitigation either ON or OFF.
 | 
						|
 | 
						|
            ON:  More then 1 pkt received (per intr.) OR we are dropping
 | 
						|
             OFF: Only 1 pkt received
 | 
						|
 | 
						|
             Note. We only use min and max (0, 15) settings from mit_table */
 | 
						|
 | 
						|
 | 
						|
          if( tp->flags &  HAS_INTR_MITIGATION) {
 | 
						|
                 if( received > 1 ) {
 | 
						|
                         if( ! tp->mit_on ) {
 | 
						|
                                 tp->mit_on = 1;
 | 
						|
                                 iowrite32(mit_table[MIT_TABLE], tp->base_addr + CSR11);
 | 
						|
                         }
 | 
						|
                  }
 | 
						|
                 else {
 | 
						|
                         if( tp->mit_on ) {
 | 
						|
                                 tp->mit_on = 0;
 | 
						|
                                 iowrite32(0, tp->base_addr + CSR11);
 | 
						|
                         }
 | 
						|
                  }
 | 
						|
          }
 | 
						|
 | 
						|
#endif /* CONFIG_TULIP_NAPI_HW_MITIGATION */
 | 
						|
 | 
						|
         tulip_refill_rx(dev);
 | 
						|
 | 
						|
         /* If RX ring is not full we are out of memory. */
 | 
						|
         if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL)
 | 
						|
		 goto oom;
 | 
						|
 | 
						|
         /* Remove us from polling list and enable RX intr. */
 | 
						|
 | 
						|
         napi_complete(napi);
 | 
						|
         iowrite32(tulip_tbl[tp->chip_id].valid_intrs, tp->base_addr+CSR7);
 | 
						|
 | 
						|
         /* The last op happens after poll completion. Which means the following:
 | 
						|
          * 1. it can race with disabling irqs in irq handler
 | 
						|
          * 2. it can race with dise/enabling irqs in other poll threads
 | 
						|
          * 3. if an irq raised after beginning loop, it will be immediately
 | 
						|
          *    triggered here.
 | 
						|
          *
 | 
						|
          * Summarizing: the logic results in some redundant irqs both
 | 
						|
          * due to races in masking and due to too late acking of already
 | 
						|
          * processed irqs. But it must not result in losing events.
 | 
						|
          */
 | 
						|
 | 
						|
         return work_done;
 | 
						|
 | 
						|
 not_done:
 | 
						|
         if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/2 ||
 | 
						|
             tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL)
 | 
						|
                 tulip_refill_rx(dev);
 | 
						|
 | 
						|
         if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL)
 | 
						|
		 goto oom;
 | 
						|
 | 
						|
         return work_done;
 | 
						|
 | 
						|
 oom:    /* Executed with RX ints disabled */
 | 
						|
 | 
						|
         /* Start timer, stop polling, but do not enable rx interrupts. */
 | 
						|
         mod_timer(&tp->oom_timer, jiffies+1);
 | 
						|
 | 
						|
         /* Think: timer_pending() was an explicit signature of bug.
 | 
						|
          * Timer can be pending now but fired and completed
 | 
						|
          * before we did napi_complete(). See? We would lose it. */
 | 
						|
 | 
						|
         /* remove ourselves from the polling list */
 | 
						|
         napi_complete(napi);
 | 
						|
 | 
						|
         return work_done;
 | 
						|
}
 | 
						|
 | 
						|
#else /* CONFIG_TULIP_NAPI */
 | 
						|
 | 
						|
static int tulip_rx(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct tulip_private *tp = netdev_priv(dev);
 | 
						|
	int entry = tp->cur_rx % RX_RING_SIZE;
 | 
						|
	int rx_work_limit = tp->dirty_rx + RX_RING_SIZE - tp->cur_rx;
 | 
						|
	int received = 0;
 | 
						|
 | 
						|
	if (tulip_debug > 4)
 | 
						|
		netdev_dbg(dev, "In tulip_rx(), entry %d %08x\n",
 | 
						|
			   entry, tp->rx_ring[entry].status);
 | 
						|
	/* If we own the next entry, it is a new packet. Send it up. */
 | 
						|
	while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) {
 | 
						|
		s32 status = le32_to_cpu(tp->rx_ring[entry].status);
 | 
						|
		short pkt_len;
 | 
						|
 | 
						|
		if (tulip_debug > 5)
 | 
						|
			netdev_dbg(dev, "In tulip_rx(), entry %d %08x\n",
 | 
						|
				   entry, status);
 | 
						|
		if (--rx_work_limit < 0)
 | 
						|
			break;
 | 
						|
 | 
						|
		/*
 | 
						|
		  Omit the four octet CRC from the length.
 | 
						|
		  (May not be considered valid until we have
 | 
						|
		  checked status for RxLengthOver2047 bits)
 | 
						|
		*/
 | 
						|
		pkt_len = ((status >> 16) & 0x7ff) - 4;
 | 
						|
		/*
 | 
						|
		  Maximum pkt_len is 1518 (1514 + vlan header)
 | 
						|
		  Anything higher than this is always invalid
 | 
						|
		  regardless of RxLengthOver2047 bits
 | 
						|
		*/
 | 
						|
 | 
						|
		if ((status & (RxLengthOver2047 |
 | 
						|
			       RxDescCRCError |
 | 
						|
			       RxDescCollisionSeen |
 | 
						|
			       RxDescRunt |
 | 
						|
			       RxDescDescErr |
 | 
						|
			       RxWholePkt))        != RxWholePkt ||
 | 
						|
		    pkt_len > 1518) {
 | 
						|
			if ((status & (RxLengthOver2047 |
 | 
						|
			     RxWholePkt))         != RxWholePkt) {
 | 
						|
				/* Ingore earlier buffers. */
 | 
						|
				if ((status & 0xffff) != 0x7fff) {
 | 
						|
					if (tulip_debug > 1)
 | 
						|
						netdev_warn(dev,
 | 
						|
							    "Oversized Ethernet frame spanned multiple buffers, status %08x!\n",
 | 
						|
							    status);
 | 
						|
					dev->stats.rx_length_errors++;
 | 
						|
				}
 | 
						|
			} else {
 | 
						|
				/* There was a fatal error. */
 | 
						|
				if (tulip_debug > 2)
 | 
						|
					netdev_dbg(dev, "Receive error, Rx status %08x\n",
 | 
						|
						   status);
 | 
						|
				dev->stats.rx_errors++; /* end of a packet.*/
 | 
						|
				if (pkt_len > 1518 ||
 | 
						|
				    (status & RxDescRunt))
 | 
						|
					dev->stats.rx_length_errors++;
 | 
						|
				if (status & 0x0004)
 | 
						|
					dev->stats.rx_frame_errors++;
 | 
						|
				if (status & 0x0002)
 | 
						|
					dev->stats.rx_crc_errors++;
 | 
						|
				if (status & 0x0001)
 | 
						|
					dev->stats.rx_fifo_errors++;
 | 
						|
			}
 | 
						|
		} else {
 | 
						|
			struct sk_buff *skb;
 | 
						|
 | 
						|
			/* Check if the packet is long enough to accept without copying
 | 
						|
			   to a minimally-sized skbuff. */
 | 
						|
			if (pkt_len < tulip_rx_copybreak &&
 | 
						|
			    (skb = netdev_alloc_skb(dev, pkt_len + 2)) != NULL) {
 | 
						|
				skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
						|
				pci_dma_sync_single_for_cpu(tp->pdev,
 | 
						|
							    tp->rx_buffers[entry].mapping,
 | 
						|
							    pkt_len, PCI_DMA_FROMDEVICE);
 | 
						|
#if ! defined(__alpha__)
 | 
						|
				skb_copy_to_linear_data(skb, tp->rx_buffers[entry].skb->data,
 | 
						|
						 pkt_len);
 | 
						|
				skb_put(skb, pkt_len);
 | 
						|
#else
 | 
						|
				memcpy(skb_put(skb, pkt_len),
 | 
						|
				       tp->rx_buffers[entry].skb->data,
 | 
						|
				       pkt_len);
 | 
						|
#endif
 | 
						|
				pci_dma_sync_single_for_device(tp->pdev,
 | 
						|
							       tp->rx_buffers[entry].mapping,
 | 
						|
							       pkt_len, PCI_DMA_FROMDEVICE);
 | 
						|
			} else { 	/* Pass up the skb already on the Rx ring. */
 | 
						|
				char *temp = skb_put(skb = tp->rx_buffers[entry].skb,
 | 
						|
						     pkt_len);
 | 
						|
 | 
						|
#ifndef final_version
 | 
						|
				if (tp->rx_buffers[entry].mapping !=
 | 
						|
				    le32_to_cpu(tp->rx_ring[entry].buffer1)) {
 | 
						|
					dev_err(&dev->dev,
 | 
						|
						"Internal fault: The skbuff addresses do not match in tulip_rx: %08x vs. %Lx %p / %p\n",
 | 
						|
						le32_to_cpu(tp->rx_ring[entry].buffer1),
 | 
						|
						(long long)tp->rx_buffers[entry].mapping,
 | 
						|
						skb->head, temp);
 | 
						|
				}
 | 
						|
#endif
 | 
						|
 | 
						|
				pci_unmap_single(tp->pdev, tp->rx_buffers[entry].mapping,
 | 
						|
						 PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
 | 
						|
 | 
						|
				tp->rx_buffers[entry].skb = NULL;
 | 
						|
				tp->rx_buffers[entry].mapping = 0;
 | 
						|
			}
 | 
						|
			skb->protocol = eth_type_trans(skb, dev);
 | 
						|
 | 
						|
			netif_rx(skb);
 | 
						|
 | 
						|
			dev->stats.rx_packets++;
 | 
						|
			dev->stats.rx_bytes += pkt_len;
 | 
						|
		}
 | 
						|
		received++;
 | 
						|
		entry = (++tp->cur_rx) % RX_RING_SIZE;
 | 
						|
	}
 | 
						|
	return received;
 | 
						|
}
 | 
						|
#endif  /* CONFIG_TULIP_NAPI */
 | 
						|
 | 
						|
static inline unsigned int phy_interrupt (struct net_device *dev)
 | 
						|
{
 | 
						|
#ifdef __hppa__
 | 
						|
	struct tulip_private *tp = netdev_priv(dev);
 | 
						|
	int csr12 = ioread32(tp->base_addr + CSR12) & 0xff;
 | 
						|
 | 
						|
	if (csr12 != tp->csr12_shadow) {
 | 
						|
		/* ack interrupt */
 | 
						|
		iowrite32(csr12 | 0x02, tp->base_addr + CSR12);
 | 
						|
		tp->csr12_shadow = csr12;
 | 
						|
		/* do link change stuff */
 | 
						|
		spin_lock(&tp->lock);
 | 
						|
		tulip_check_duplex(dev);
 | 
						|
		spin_unlock(&tp->lock);
 | 
						|
		/* clear irq ack bit */
 | 
						|
		iowrite32(csr12 & ~0x02, tp->base_addr + CSR12);
 | 
						|
 | 
						|
		return 1;
 | 
						|
	}
 | 
						|
#endif
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
/* The interrupt handler does all of the Rx thread work and cleans up
 | 
						|
   after the Tx thread. */
 | 
						|
irqreturn_t tulip_interrupt(int irq, void *dev_instance)
 | 
						|
{
 | 
						|
	struct net_device *dev = (struct net_device *)dev_instance;
 | 
						|
	struct tulip_private *tp = netdev_priv(dev);
 | 
						|
	void __iomem *ioaddr = tp->base_addr;
 | 
						|
	int csr5;
 | 
						|
	int missed;
 | 
						|
	int rx = 0;
 | 
						|
	int tx = 0;
 | 
						|
	int oi = 0;
 | 
						|
	int maxrx = RX_RING_SIZE;
 | 
						|
	int maxtx = TX_RING_SIZE;
 | 
						|
	int maxoi = TX_RING_SIZE;
 | 
						|
#ifdef CONFIG_TULIP_NAPI
 | 
						|
	int rxd = 0;
 | 
						|
#else
 | 
						|
	int entry;
 | 
						|
#endif
 | 
						|
	unsigned int work_count = tulip_max_interrupt_work;
 | 
						|
	unsigned int handled = 0;
 | 
						|
 | 
						|
	/* Let's see whether the interrupt really is for us */
 | 
						|
	csr5 = ioread32(ioaddr + CSR5);
 | 
						|
 | 
						|
        if (tp->flags & HAS_PHY_IRQ)
 | 
						|
	        handled = phy_interrupt (dev);
 | 
						|
 | 
						|
	if ((csr5 & (NormalIntr|AbnormalIntr)) == 0)
 | 
						|
		return IRQ_RETVAL(handled);
 | 
						|
 | 
						|
	tp->nir++;
 | 
						|
 | 
						|
	do {
 | 
						|
 | 
						|
#ifdef CONFIG_TULIP_NAPI
 | 
						|
 | 
						|
		if (!rxd && (csr5 & (RxIntr | RxNoBuf))) {
 | 
						|
			rxd++;
 | 
						|
			/* Mask RX intrs and add the device to poll list. */
 | 
						|
			iowrite32(tulip_tbl[tp->chip_id].valid_intrs&~RxPollInt, ioaddr + CSR7);
 | 
						|
			napi_schedule(&tp->napi);
 | 
						|
 | 
						|
			if (!(csr5&~(AbnormalIntr|NormalIntr|RxPollInt|TPLnkPass)))
 | 
						|
                               break;
 | 
						|
		}
 | 
						|
 | 
						|
               /* Acknowledge the interrupt sources we handle here ASAP
 | 
						|
                  the poll function does Rx and RxNoBuf acking */
 | 
						|
 | 
						|
		iowrite32(csr5 & 0x0001ff3f, ioaddr + CSR5);
 | 
						|
 | 
						|
#else
 | 
						|
		/* Acknowledge all of the current interrupt sources ASAP. */
 | 
						|
		iowrite32(csr5 & 0x0001ffff, ioaddr + CSR5);
 | 
						|
 | 
						|
 | 
						|
		if (csr5 & (RxIntr | RxNoBuf)) {
 | 
						|
				rx += tulip_rx(dev);
 | 
						|
			tulip_refill_rx(dev);
 | 
						|
		}
 | 
						|
 | 
						|
#endif /*  CONFIG_TULIP_NAPI */
 | 
						|
 | 
						|
		if (tulip_debug > 4)
 | 
						|
			netdev_dbg(dev, "interrupt  csr5=%#8.8x new csr5=%#8.8x\n",
 | 
						|
				   csr5, ioread32(ioaddr + CSR5));
 | 
						|
 | 
						|
 | 
						|
		if (csr5 & (TxNoBuf | TxDied | TxIntr | TimerInt)) {
 | 
						|
			unsigned int dirty_tx;
 | 
						|
 | 
						|
			spin_lock(&tp->lock);
 | 
						|
 | 
						|
			for (dirty_tx = tp->dirty_tx; tp->cur_tx - dirty_tx > 0;
 | 
						|
				 dirty_tx++) {
 | 
						|
				int entry = dirty_tx % TX_RING_SIZE;
 | 
						|
				int status = le32_to_cpu(tp->tx_ring[entry].status);
 | 
						|
 | 
						|
				if (status < 0)
 | 
						|
					break;			/* It still has not been Txed */
 | 
						|
 | 
						|
				/* Check for Rx filter setup frames. */
 | 
						|
				if (tp->tx_buffers[entry].skb == NULL) {
 | 
						|
					/* test because dummy frames not mapped */
 | 
						|
					if (tp->tx_buffers[entry].mapping)
 | 
						|
						pci_unmap_single(tp->pdev,
 | 
						|
							 tp->tx_buffers[entry].mapping,
 | 
						|
							 sizeof(tp->setup_frame),
 | 
						|
							 PCI_DMA_TODEVICE);
 | 
						|
					continue;
 | 
						|
				}
 | 
						|
 | 
						|
				if (status & 0x8000) {
 | 
						|
					/* There was an major error, log it. */
 | 
						|
#ifndef final_version
 | 
						|
					if (tulip_debug > 1)
 | 
						|
						netdev_dbg(dev, "Transmit error, Tx status %08x\n",
 | 
						|
							   status);
 | 
						|
#endif
 | 
						|
					dev->stats.tx_errors++;
 | 
						|
					if (status & 0x4104)
 | 
						|
						dev->stats.tx_aborted_errors++;
 | 
						|
					if (status & 0x0C00)
 | 
						|
						dev->stats.tx_carrier_errors++;
 | 
						|
					if (status & 0x0200)
 | 
						|
						dev->stats.tx_window_errors++;
 | 
						|
					if (status & 0x0002)
 | 
						|
						dev->stats.tx_fifo_errors++;
 | 
						|
					if ((status & 0x0080) && tp->full_duplex == 0)
 | 
						|
						dev->stats.tx_heartbeat_errors++;
 | 
						|
				} else {
 | 
						|
					dev->stats.tx_bytes +=
 | 
						|
						tp->tx_buffers[entry].skb->len;
 | 
						|
					dev->stats.collisions += (status >> 3) & 15;
 | 
						|
					dev->stats.tx_packets++;
 | 
						|
				}
 | 
						|
 | 
						|
				pci_unmap_single(tp->pdev, tp->tx_buffers[entry].mapping,
 | 
						|
						 tp->tx_buffers[entry].skb->len,
 | 
						|
						 PCI_DMA_TODEVICE);
 | 
						|
 | 
						|
				/* Free the original skb. */
 | 
						|
				dev_kfree_skb_irq(tp->tx_buffers[entry].skb);
 | 
						|
				tp->tx_buffers[entry].skb = NULL;
 | 
						|
				tp->tx_buffers[entry].mapping = 0;
 | 
						|
				tx++;
 | 
						|
			}
 | 
						|
 | 
						|
#ifndef final_version
 | 
						|
			if (tp->cur_tx - dirty_tx > TX_RING_SIZE) {
 | 
						|
				dev_err(&dev->dev,
 | 
						|
					"Out-of-sync dirty pointer, %d vs. %d\n",
 | 
						|
					dirty_tx, tp->cur_tx);
 | 
						|
				dirty_tx += TX_RING_SIZE;
 | 
						|
			}
 | 
						|
#endif
 | 
						|
 | 
						|
			if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2)
 | 
						|
				netif_wake_queue(dev);
 | 
						|
 | 
						|
			tp->dirty_tx = dirty_tx;
 | 
						|
			if (csr5 & TxDied) {
 | 
						|
				if (tulip_debug > 2)
 | 
						|
					dev_warn(&dev->dev,
 | 
						|
						 "The transmitter stopped.  CSR5 is %x, CSR6 %x, new CSR6 %x\n",
 | 
						|
						 csr5, ioread32(ioaddr + CSR6),
 | 
						|
						 tp->csr6);
 | 
						|
				tulip_restart_rxtx(tp);
 | 
						|
			}
 | 
						|
			spin_unlock(&tp->lock);
 | 
						|
		}
 | 
						|
 | 
						|
		/* Log errors. */
 | 
						|
		if (csr5 & AbnormalIntr) {	/* Abnormal error summary bit. */
 | 
						|
			if (csr5 == 0xffffffff)
 | 
						|
				break;
 | 
						|
			if (csr5 & TxJabber)
 | 
						|
				dev->stats.tx_errors++;
 | 
						|
			if (csr5 & TxFIFOUnderflow) {
 | 
						|
				if ((tp->csr6 & 0xC000) != 0xC000)
 | 
						|
					tp->csr6 += 0x4000;	/* Bump up the Tx threshold */
 | 
						|
				else
 | 
						|
					tp->csr6 |= 0x00200000;  /* Store-n-forward. */
 | 
						|
				/* Restart the transmit process. */
 | 
						|
				tulip_restart_rxtx(tp);
 | 
						|
				iowrite32(0, ioaddr + CSR1);
 | 
						|
			}
 | 
						|
			if (csr5 & (RxDied | RxNoBuf)) {
 | 
						|
				if (tp->flags & COMET_MAC_ADDR) {
 | 
						|
					iowrite32(tp->mc_filter[0], ioaddr + 0xAC);
 | 
						|
					iowrite32(tp->mc_filter[1], ioaddr + 0xB0);
 | 
						|
				}
 | 
						|
			}
 | 
						|
			if (csr5 & RxDied) {		/* Missed a Rx frame. */
 | 
						|
				dev->stats.rx_missed_errors += ioread32(ioaddr + CSR8) & 0xffff;
 | 
						|
				dev->stats.rx_errors++;
 | 
						|
				tulip_start_rxtx(tp);
 | 
						|
			}
 | 
						|
			/*
 | 
						|
			 * NB: t21142_lnk_change() does a del_timer_sync(), so be careful if this
 | 
						|
			 * call is ever done under the spinlock
 | 
						|
			 */
 | 
						|
			if (csr5 & (TPLnkPass | TPLnkFail | 0x08000000)) {
 | 
						|
				if (tp->link_change)
 | 
						|
					(tp->link_change)(dev, csr5);
 | 
						|
			}
 | 
						|
			if (csr5 & SystemError) {
 | 
						|
				int error = (csr5 >> 23) & 7;
 | 
						|
				/* oops, we hit a PCI error.  The code produced corresponds
 | 
						|
				 * to the reason:
 | 
						|
				 *  0 - parity error
 | 
						|
				 *  1 - master abort
 | 
						|
				 *  2 - target abort
 | 
						|
				 * Note that on parity error, we should do a software reset
 | 
						|
				 * of the chip to get it back into a sane state (according
 | 
						|
				 * to the 21142/3 docs that is).
 | 
						|
				 *   -- rmk
 | 
						|
				 */
 | 
						|
				dev_err(&dev->dev,
 | 
						|
					"(%lu) System Error occurred (%d)\n",
 | 
						|
					tp->nir, error);
 | 
						|
			}
 | 
						|
			/* Clear all error sources, included undocumented ones! */
 | 
						|
			iowrite32(0x0800f7ba, ioaddr + CSR5);
 | 
						|
			oi++;
 | 
						|
		}
 | 
						|
		if (csr5 & TimerInt) {
 | 
						|
 | 
						|
			if (tulip_debug > 2)
 | 
						|
				dev_err(&dev->dev,
 | 
						|
					"Re-enabling interrupts, %08x\n",
 | 
						|
					csr5);
 | 
						|
			iowrite32(tulip_tbl[tp->chip_id].valid_intrs, ioaddr + CSR7);
 | 
						|
			tp->ttimer = 0;
 | 
						|
			oi++;
 | 
						|
		}
 | 
						|
		if (tx > maxtx || rx > maxrx || oi > maxoi) {
 | 
						|
			if (tulip_debug > 1)
 | 
						|
				dev_warn(&dev->dev, "Too much work during an interrupt, csr5=0x%08x. (%lu) (%d,%d,%d)\n",
 | 
						|
					 csr5, tp->nir, tx, rx, oi);
 | 
						|
 | 
						|
                       /* Acknowledge all interrupt sources. */
 | 
						|
                        iowrite32(0x8001ffff, ioaddr + CSR5);
 | 
						|
                        if (tp->flags & HAS_INTR_MITIGATION) {
 | 
						|
                     /* Josip Loncaric at ICASE did extensive experimentation
 | 
						|
			to develop a good interrupt mitigation setting.*/
 | 
						|
                                iowrite32(0x8b240000, ioaddr + CSR11);
 | 
						|
                        } else if (tp->chip_id == LC82C168) {
 | 
						|
				/* the LC82C168 doesn't have a hw timer.*/
 | 
						|
				iowrite32(0x00, ioaddr + CSR7);
 | 
						|
				mod_timer(&tp->timer, RUN_AT(HZ/50));
 | 
						|
			} else {
 | 
						|
                          /* Mask all interrupting sources, set timer to
 | 
						|
				re-enable. */
 | 
						|
                                iowrite32(((~csr5) & 0x0001ebef) | AbnormalIntr | TimerInt, ioaddr + CSR7);
 | 
						|
                                iowrite32(0x0012, ioaddr + CSR11);
 | 
						|
                        }
 | 
						|
			break;
 | 
						|
		}
 | 
						|
 | 
						|
		work_count--;
 | 
						|
		if (work_count == 0)
 | 
						|
			break;
 | 
						|
 | 
						|
		csr5 = ioread32(ioaddr + CSR5);
 | 
						|
 | 
						|
#ifdef CONFIG_TULIP_NAPI
 | 
						|
		if (rxd)
 | 
						|
			csr5 &= ~RxPollInt;
 | 
						|
	} while ((csr5 & (TxNoBuf |
 | 
						|
			  TxDied |
 | 
						|
			  TxIntr |
 | 
						|
			  TimerInt |
 | 
						|
			  /* Abnormal intr. */
 | 
						|
			  RxDied |
 | 
						|
			  TxFIFOUnderflow |
 | 
						|
			  TxJabber |
 | 
						|
			  TPLnkFail |
 | 
						|
			  SystemError )) != 0);
 | 
						|
#else
 | 
						|
	} while ((csr5 & (NormalIntr|AbnormalIntr)) != 0);
 | 
						|
 | 
						|
	tulip_refill_rx(dev);
 | 
						|
 | 
						|
	/* check if the card is in suspend mode */
 | 
						|
	entry = tp->dirty_rx % RX_RING_SIZE;
 | 
						|
	if (tp->rx_buffers[entry].skb == NULL) {
 | 
						|
		if (tulip_debug > 1)
 | 
						|
			dev_warn(&dev->dev,
 | 
						|
				 "in rx suspend mode: (%lu) (tp->cur_rx = %u, ttimer = %d, rx = %d) go/stay in suspend mode\n",
 | 
						|
				 tp->nir, tp->cur_rx, tp->ttimer, rx);
 | 
						|
		if (tp->chip_id == LC82C168) {
 | 
						|
			iowrite32(0x00, ioaddr + CSR7);
 | 
						|
			mod_timer(&tp->timer, RUN_AT(HZ/50));
 | 
						|
		} else {
 | 
						|
			if (tp->ttimer == 0 || (ioread32(ioaddr + CSR11) & 0xffff) == 0) {
 | 
						|
				if (tulip_debug > 1)
 | 
						|
					dev_warn(&dev->dev,
 | 
						|
						 "in rx suspend mode: (%lu) set timer\n",
 | 
						|
						 tp->nir);
 | 
						|
				iowrite32(tulip_tbl[tp->chip_id].valid_intrs | TimerInt,
 | 
						|
					ioaddr + CSR7);
 | 
						|
				iowrite32(TimerInt, ioaddr + CSR5);
 | 
						|
				iowrite32(12, ioaddr + CSR11);
 | 
						|
				tp->ttimer = 1;
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
#endif /* CONFIG_TULIP_NAPI */
 | 
						|
 | 
						|
	if ((missed = ioread32(ioaddr + CSR8) & 0x1ffff)) {
 | 
						|
		dev->stats.rx_dropped += missed & 0x10000 ? 0x10000 : missed;
 | 
						|
	}
 | 
						|
 | 
						|
	if (tulip_debug > 4)
 | 
						|
		netdev_dbg(dev, "exiting interrupt, csr5=%#04x\n",
 | 
						|
			   ioread32(ioaddr + CSR5));
 | 
						|
 | 
						|
	return IRQ_HANDLED;
 | 
						|
}
 |