netdev: ethernet dev_alloc_skb to netdev_alloc_skb
Replaced deprecating dev_alloc_skb with netdev_alloc_skb in drivers/net/ethernet - Removed extra skb->dev = dev after netdev_alloc_skb Signed-off-by: Pradeep A Dalvi <netdev@pradeepdalvi.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
		
					parent
					
						
							
								1d26643054
							
						
					
				
			
			
				commit
				
					
						21a4e46995
					
				
			
		
					 21 changed files with 46 additions and 51 deletions
				
			
		| 
						 | 
					@ -397,7 +397,7 @@ static int macb_rx_frame(struct macb *bp, unsigned int first_frag,
 | 
				
			||||||
	netdev_dbg(bp->dev, "macb_rx_frame frags %u - %u (len %u)\n",
 | 
						netdev_dbg(bp->dev, "macb_rx_frame frags %u - %u (len %u)\n",
 | 
				
			||||||
		   first_frag, last_frag, len);
 | 
							   first_frag, last_frag, len);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	skb = dev_alloc_skb(len + RX_OFFSET);
 | 
						skb = netdev_alloc_skb(bp->dev, len + RX_OFFSET);
 | 
				
			||||||
	if (!skb) {
 | 
						if (!skb) {
 | 
				
			||||||
		bp->stats.rx_dropped++;
 | 
							bp->stats.rx_dropped++;
 | 
				
			||||||
		for (frag = first_frag; ; frag = NEXT_RX(frag)) {
 | 
							for (frag = first_frag; ; frag = NEXT_RX(frag)) {
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -911,7 +911,7 @@ dma_rx(struct net_device *dev)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Malloc up new buffer. */
 | 
						/* Malloc up new buffer. */
 | 
				
			||||||
	skb = dev_alloc_skb(length + 2);
 | 
						skb = netdev_alloc_skb(dev, length + 2);
 | 
				
			||||||
	if (skb == NULL) {
 | 
						if (skb == NULL) {
 | 
				
			||||||
		if (net_debug)	/* I don't think we want to do this to a stressed system */
 | 
							if (net_debug)	/* I don't think we want to do this to a stressed system */
 | 
				
			||||||
			printk("%s: Memory squeeze, dropping packet.\n", dev->name);
 | 
								printk("%s: Memory squeeze, dropping packet.\n", dev->name);
 | 
				
			||||||
| 
						 | 
					@ -1616,7 +1616,7 @@ net_rx(struct net_device *dev)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Malloc up new buffer. */
 | 
						/* Malloc up new buffer. */
 | 
				
			||||||
	skb = dev_alloc_skb(length + 2);
 | 
						skb = netdev_alloc_skb(dev, length + 2);
 | 
				
			||||||
	if (skb == NULL) {
 | 
						if (skb == NULL) {
 | 
				
			||||||
#if 0		/* Again, this seems a cruel thing to do */
 | 
					#if 0		/* Again, this seems a cruel thing to do */
 | 
				
			||||||
		printk(KERN_WARNING "%s: Memory squeeze, dropping packet.\n", dev->name);
 | 
							printk(KERN_WARNING "%s: Memory squeeze, dropping packet.\n", dev->name);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -282,7 +282,7 @@ static int ep93xx_rx(struct net_device *dev, int processed, int budget)
 | 
				
			||||||
		if (rstat0 & RSTAT0_CRCI)
 | 
							if (rstat0 & RSTAT0_CRCI)
 | 
				
			||||||
			length -= 4;
 | 
								length -= 4;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		skb = dev_alloc_skb(length + 2);
 | 
							skb = netdev_alloc_skb(dev, length + 2);
 | 
				
			||||||
		if (likely(skb != NULL)) {
 | 
							if (likely(skb != NULL)) {
 | 
				
			||||||
			struct ep93xx_rdesc *rxd = &ep->descs->rdesc[entry];
 | 
								struct ep93xx_rdesc *rxd = &ep->descs->rdesc[entry];
 | 
				
			||||||
			skb_reserve(skb, 2);
 | 
								skb_reserve(skb, 2);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1028,7 +1028,7 @@ dm9000_rx(struct net_device *dev)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* Move data from DM9000 */
 | 
							/* Move data from DM9000 */
 | 
				
			||||||
		if (GoodPacket &&
 | 
							if (GoodPacket &&
 | 
				
			||||||
		    ((skb = dev_alloc_skb(RxLen + 4)) != NULL)) {
 | 
							    ((skb = netdev_alloc_skb(dev, RxLen + 4)) != NULL)) {
 | 
				
			||||||
			skb_reserve(skb, 2);
 | 
								skb_reserve(skb, 2);
 | 
				
			||||||
			rdptr = (u8 *) skb_put(skb, RxLen - 4);
 | 
								rdptr = (u8 *) skb_put(skb, RxLen - 4);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -986,8 +986,10 @@ static int ewrk3_rx(struct net_device *dev)
 | 
				
			||||||
						dev->stats.rx_fifo_errors++;
 | 
											dev->stats.rx_fifo_errors++;
 | 
				
			||||||
				} else {
 | 
									} else {
 | 
				
			||||||
					struct sk_buff *skb;
 | 
										struct sk_buff *skb;
 | 
				
			||||||
 | 
										skb = netdev_alloc_skb(dev,
 | 
				
			||||||
 | 
												pkt_len + 2);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
					if ((skb = dev_alloc_skb(pkt_len + 2)) != NULL) {
 | 
										if (skb != NULL) {
 | 
				
			||||||
						unsigned char *p;
 | 
											unsigned char *p;
 | 
				
			||||||
						skb_reserve(skb, 2);	/* Align to 16 bytes */
 | 
											skb_reserve(skb, 2);	/* Align to 16 bytes */
 | 
				
			||||||
						p = skb_put(skb, pkt_len);
 | 
											p = skb_put(skb, pkt_len);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -439,7 +439,7 @@ static void de_rx (struct de_private *de)
 | 
				
			||||||
			  rx_tail, status, len, copying_skb);
 | 
								  rx_tail, status, len, copying_skb);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		buflen = copying_skb ? (len + RX_OFFSET) : de->rx_buf_sz;
 | 
							buflen = copying_skb ? (len + RX_OFFSET) : de->rx_buf_sz;
 | 
				
			||||||
		copy_skb = dev_alloc_skb (buflen);
 | 
							copy_skb = netdev_alloc_skb(de->dev, buflen);
 | 
				
			||||||
		if (unlikely(!copy_skb)) {
 | 
							if (unlikely(!copy_skb)) {
 | 
				
			||||||
			de->net_stats.rx_dropped++;
 | 
								de->net_stats.rx_dropped++;
 | 
				
			||||||
			drop = 1;
 | 
								drop = 1;
 | 
				
			||||||
| 
						 | 
					@ -1283,12 +1283,10 @@ static int de_refill_rx (struct de_private *de)
 | 
				
			||||||
	for (i = 0; i < DE_RX_RING_SIZE; i++) {
 | 
						for (i = 0; i < DE_RX_RING_SIZE; i++) {
 | 
				
			||||||
		struct sk_buff *skb;
 | 
							struct sk_buff *skb;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		skb = dev_alloc_skb(de->rx_buf_sz);
 | 
							skb = netdev_alloc_skb(de->dev, de->rx_buf_sz);
 | 
				
			||||||
		if (!skb)
 | 
							if (!skb)
 | 
				
			||||||
			goto err_out;
 | 
								goto err_out;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		skb->dev = de->dev;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		de->rx_skb[i].mapping = pci_map_single(de->pdev,
 | 
							de->rx_skb[i].mapping = pci_map_single(de->pdev,
 | 
				
			||||||
			skb->data, de->rx_buf_sz, PCI_DMA_FROMDEVICE);
 | 
								skb->data, de->rx_buf_sz, PCI_DMA_FROMDEVICE);
 | 
				
			||||||
		de->rx_skb[i].skb = skb;
 | 
							de->rx_skb[i].skb = skb;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3598,7 +3598,7 @@ de4x5_alloc_rx_buff(struct net_device *dev, int index, int len)
 | 
				
			||||||
    struct sk_buff *ret;
 | 
					    struct sk_buff *ret;
 | 
				
			||||||
    u_long i=0, tmp;
 | 
					    u_long i=0, tmp;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    p = dev_alloc_skb(IEEE802_3_SZ + DE4X5_ALIGN + 2);
 | 
					    p = netdev_alloc_skb(dev, IEEE802_3_SZ + DE4X5_ALIGN + 2);
 | 
				
			||||||
    if (!p) return NULL;
 | 
					    if (!p) return NULL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    tmp = virt_to_bus(p->data);
 | 
					    tmp = virt_to_bus(p->data);
 | 
				
			||||||
| 
						 | 
					@ -3618,7 +3618,7 @@ de4x5_alloc_rx_buff(struct net_device *dev, int index, int len)
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
    if (lp->state != OPEN) return (struct sk_buff *)1; /* Fake out the open */
 | 
					    if (lp->state != OPEN) return (struct sk_buff *)1; /* Fake out the open */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    p = dev_alloc_skb(len + 2);
 | 
					    p = netdev_alloc_skb(dev, len + 2);
 | 
				
			||||||
    if (!p) return NULL;
 | 
					    if (!p) return NULL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    skb_reserve(p, 2);	                               /* Align */
 | 
					    skb_reserve(p, 2);	                               /* Align */
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -69,7 +69,8 @@ int tulip_refill_rx(struct net_device *dev)
 | 
				
			||||||
			struct sk_buff *skb;
 | 
								struct sk_buff *skb;
 | 
				
			||||||
			dma_addr_t mapping;
 | 
								dma_addr_t mapping;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			skb = tp->rx_buffers[entry].skb = dev_alloc_skb(PKT_BUF_SZ);
 | 
								skb = tp->rx_buffers[entry].skb =
 | 
				
			||||||
 | 
									netdev_alloc_skb(dev, PKT_BUF_SZ);
 | 
				
			||||||
			if (skb == NULL)
 | 
								if (skb == NULL)
 | 
				
			||||||
				break;
 | 
									break;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -77,7 +78,6 @@ int tulip_refill_rx(struct net_device *dev)
 | 
				
			||||||
						 PCI_DMA_FROMDEVICE);
 | 
											 PCI_DMA_FROMDEVICE);
 | 
				
			||||||
			tp->rx_buffers[entry].mapping = mapping;
 | 
								tp->rx_buffers[entry].mapping = mapping;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			skb->dev = dev;			/* Mark as being used by this device. */
 | 
					 | 
				
			||||||
			tp->rx_ring[entry].buffer1 = cpu_to_le32(mapping);
 | 
								tp->rx_ring[entry].buffer1 = cpu_to_le32(mapping);
 | 
				
			||||||
			refilled++;
 | 
								refilled++;
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
| 
						 | 
					@ -202,7 +202,7 @@ int tulip_poll(struct napi_struct *napi, int budget)
 | 
				
			||||||
                               /* Check if the packet is long enough to accept without copying
 | 
					                               /* Check if the packet is long enough to accept without copying
 | 
				
			||||||
                                  to a minimally-sized skbuff. */
 | 
					                                  to a minimally-sized skbuff. */
 | 
				
			||||||
                               if (pkt_len < tulip_rx_copybreak &&
 | 
					                               if (pkt_len < tulip_rx_copybreak &&
 | 
				
			||||||
                                   (skb = dev_alloc_skb(pkt_len + 2)) != NULL) {
 | 
					                                   (skb = netdev_alloc_skb(dev, pkt_len + 2)) != NULL) {
 | 
				
			||||||
                                       skb_reserve(skb, 2);    /* 16 byte align the IP header */
 | 
					                                       skb_reserve(skb, 2);    /* 16 byte align the IP header */
 | 
				
			||||||
                                       pci_dma_sync_single_for_cpu(tp->pdev,
 | 
					                                       pci_dma_sync_single_for_cpu(tp->pdev,
 | 
				
			||||||
								   tp->rx_buffers[entry].mapping,
 | 
													   tp->rx_buffers[entry].mapping,
 | 
				
			||||||
| 
						 | 
					@ -428,7 +428,7 @@ static int tulip_rx(struct net_device *dev)
 | 
				
			||||||
			/* Check if the packet is long enough to accept without copying
 | 
								/* Check if the packet is long enough to accept without copying
 | 
				
			||||||
			   to a minimally-sized skbuff. */
 | 
								   to a minimally-sized skbuff. */
 | 
				
			||||||
			if (pkt_len < tulip_rx_copybreak &&
 | 
								if (pkt_len < tulip_rx_copybreak &&
 | 
				
			||||||
			    (skb = dev_alloc_skb(pkt_len + 2)) != NULL) {
 | 
								    (skb = netdev_alloc_skb(dev, pkt_len + 2)) != NULL) {
 | 
				
			||||||
				skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
									skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
				
			||||||
				pci_dma_sync_single_for_cpu(tp->pdev,
 | 
									pci_dma_sync_single_for_cpu(tp->pdev,
 | 
				
			||||||
							    tp->rx_buffers[entry].mapping,
 | 
												    tp->rx_buffers[entry].mapping,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -636,16 +636,15 @@ static void tulip_init_ring(struct net_device *dev)
 | 
				
			||||||
		dma_addr_t mapping;
 | 
							dma_addr_t mapping;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* Note the receive buffer must be longword aligned.
 | 
							/* Note the receive buffer must be longword aligned.
 | 
				
			||||||
		   dev_alloc_skb() provides 16 byte alignment.  But do *not*
 | 
							   netdev_alloc_skb() provides 16 byte alignment.  But do *not*
 | 
				
			||||||
		   use skb_reserve() to align the IP header! */
 | 
							   use skb_reserve() to align the IP header! */
 | 
				
			||||||
		struct sk_buff *skb = dev_alloc_skb(PKT_BUF_SZ);
 | 
							struct sk_buff *skb = netdev_alloc_skb(dev, PKT_BUF_SZ);
 | 
				
			||||||
		tp->rx_buffers[i].skb = skb;
 | 
							tp->rx_buffers[i].skb = skb;
 | 
				
			||||||
		if (skb == NULL)
 | 
							if (skb == NULL)
 | 
				
			||||||
			break;
 | 
								break;
 | 
				
			||||||
		mapping = pci_map_single(tp->pdev, skb->data,
 | 
							mapping = pci_map_single(tp->pdev, skb->data,
 | 
				
			||||||
					 PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
 | 
										 PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
 | 
				
			||||||
		tp->rx_buffers[i].mapping = mapping;
 | 
							tp->rx_buffers[i].mapping = mapping;
 | 
				
			||||||
		skb->dev = dev;			/* Mark as being used by this device. */
 | 
					 | 
				
			||||||
		tp->rx_ring[i].status = cpu_to_le32(DescOwned);	/* Owned by Tulip chip */
 | 
							tp->rx_ring[i].status = cpu_to_le32(DescOwned);	/* Owned by Tulip chip */
 | 
				
			||||||
		tp->rx_ring[i].buffer1 = cpu_to_le32(mapping);
 | 
							tp->rx_ring[i].buffer1 = cpu_to_le32(mapping);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -815,7 +815,7 @@ static void init_rxtx_rings(struct net_device *dev)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Fill in the Rx buffers.  Handle allocation failure gracefully. */
 | 
						/* Fill in the Rx buffers.  Handle allocation failure gracefully. */
 | 
				
			||||||
	for (i = 0; i < RX_RING_SIZE; i++) {
 | 
						for (i = 0; i < RX_RING_SIZE; i++) {
 | 
				
			||||||
		struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz);
 | 
							struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz);
 | 
				
			||||||
		np->rx_skbuff[i] = skb;
 | 
							np->rx_skbuff[i] = skb;
 | 
				
			||||||
		if (skb == NULL)
 | 
							if (skb == NULL)
 | 
				
			||||||
			break;
 | 
								break;
 | 
				
			||||||
| 
						 | 
					@ -1231,7 +1231,7 @@ static int netdev_rx(struct net_device *dev)
 | 
				
			||||||
			/* Check if the packet is long enough to accept without copying
 | 
								/* Check if the packet is long enough to accept without copying
 | 
				
			||||||
			   to a minimally-sized skbuff. */
 | 
								   to a minimally-sized skbuff. */
 | 
				
			||||||
			if (pkt_len < rx_copybreak &&
 | 
								if (pkt_len < rx_copybreak &&
 | 
				
			||||||
			    (skb = dev_alloc_skb(pkt_len + 2)) != NULL) {
 | 
								    (skb = netdev_alloc_skb(dev, pkt_len + 2)) != NULL) {
 | 
				
			||||||
				skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
									skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
				
			||||||
				pci_dma_sync_single_for_cpu(np->pci_dev,np->rx_addr[entry],
 | 
									pci_dma_sync_single_for_cpu(np->pci_dev,np->rx_addr[entry],
 | 
				
			||||||
							    np->rx_skbuff[entry]->len,
 | 
												    np->rx_skbuff[entry]->len,
 | 
				
			||||||
| 
						 | 
					@ -1270,7 +1270,7 @@ static int netdev_rx(struct net_device *dev)
 | 
				
			||||||
		struct sk_buff *skb;
 | 
							struct sk_buff *skb;
 | 
				
			||||||
		entry = np->dirty_rx % RX_RING_SIZE;
 | 
							entry = np->dirty_rx % RX_RING_SIZE;
 | 
				
			||||||
		if (np->rx_skbuff[entry] == NULL) {
 | 
							if (np->rx_skbuff[entry] == NULL) {
 | 
				
			||||||
			skb = dev_alloc_skb(np->rx_buf_sz);
 | 
								skb = netdev_alloc_skb(dev, np->rx_buf_sz);
 | 
				
			||||||
			np->rx_skbuff[entry] = skb;
 | 
								np->rx_skbuff[entry] = skb;
 | 
				
			||||||
			if (skb == NULL)
 | 
								if (skb == NULL)
 | 
				
			||||||
				break;			/* Better luck next round. */
 | 
									break;			/* Better luck next round. */
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1084,7 +1084,7 @@ investigate_read_descriptor(struct net_device *dev, struct xircom_private *card,
 | 
				
			||||||
			pkt_len = 1518;
 | 
								pkt_len = 1518;
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		skb = dev_alloc_skb(pkt_len + 2);
 | 
							skb = netdev_alloc_skb(dev, pkt_len + 2);
 | 
				
			||||||
		if (skb == NULL) {
 | 
							if (skb == NULL) {
 | 
				
			||||||
			dev->stats.rx_dropped++;
 | 
								dev->stats.rx_dropped++;
 | 
				
			||||||
			goto out;
 | 
								goto out;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -335,7 +335,7 @@ static void de600_rx_intr(struct net_device *dev)
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	skb = dev_alloc_skb(size+2);
 | 
						skb = netdev_alloc_skb(dev, size + 2);
 | 
				
			||||||
	if (skb == NULL) {
 | 
						if (skb == NULL) {
 | 
				
			||||||
		printk("%s: Couldn't allocate a sk_buff of size %d.\n", dev->name, size);
 | 
							printk("%s: Couldn't allocate a sk_buff of size %d.\n", dev->name, size);
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -650,7 +650,7 @@ static int de620_rx_intr(struct net_device *dev)
 | 
				
			||||||
		printk(KERN_WARNING "%s: Illegal packet size: %d!\n", dev->name, size);
 | 
							printk(KERN_WARNING "%s: Illegal packet size: %d!\n", dev->name, size);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
	else { /* Good packet? */
 | 
						else { /* Good packet? */
 | 
				
			||||||
		skb = dev_alloc_skb(size+2);
 | 
							skb = netdev_alloc_skb(dev, size + 2);
 | 
				
			||||||
		if (skb == NULL) { /* Yeah, but no place to put it... */
 | 
							if (skb == NULL) { /* Yeah, but no place to put it... */
 | 
				
			||||||
			printk(KERN_WARNING "%s: Couldn't allocate a sk_buff of size %d.\n", dev->name, size);
 | 
								printk(KERN_WARNING "%s: Couldn't allocate a sk_buff of size %d.\n", dev->name, size);
 | 
				
			||||||
			dev->stats.rx_dropped++;
 | 
								dev->stats.rx_dropped++;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1020,11 +1020,11 @@ static void init_ring(struct net_device *dev)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Fill in the Rx buffers.  Handle allocation failure gracefully. */
 | 
						/* Fill in the Rx buffers.  Handle allocation failure gracefully. */
 | 
				
			||||||
	for (i = 0; i < RX_RING_SIZE; i++) {
 | 
						for (i = 0; i < RX_RING_SIZE; i++) {
 | 
				
			||||||
		struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz + 2);
 | 
							struct sk_buff *skb =
 | 
				
			||||||
 | 
								netdev_alloc_skb(dev, np->rx_buf_sz + 2);
 | 
				
			||||||
		np->rx_skbuff[i] = skb;
 | 
							np->rx_skbuff[i] = skb;
 | 
				
			||||||
		if (skb == NULL)
 | 
							if (skb == NULL)
 | 
				
			||||||
			break;
 | 
								break;
 | 
				
			||||||
		skb->dev = dev;		/* Mark as being used by this device. */
 | 
					 | 
				
			||||||
		skb_reserve(skb, 2);	/* 16 byte align the IP header. */
 | 
							skb_reserve(skb, 2);	/* 16 byte align the IP header. */
 | 
				
			||||||
		np->rx_ring[i].frag[0].addr = cpu_to_le32(
 | 
							np->rx_ring[i].frag[0].addr = cpu_to_le32(
 | 
				
			||||||
			dma_map_single(&np->pci_dev->dev, skb->data,
 | 
								dma_map_single(&np->pci_dev->dev, skb->data,
 | 
				
			||||||
| 
						 | 
					@ -1358,7 +1358,7 @@ static void rx_poll(unsigned long data)
 | 
				
			||||||
			/* Check if the packet is long enough to accept without copying
 | 
								/* Check if the packet is long enough to accept without copying
 | 
				
			||||||
			   to a minimally-sized skbuff. */
 | 
								   to a minimally-sized skbuff. */
 | 
				
			||||||
			if (pkt_len < rx_copybreak &&
 | 
								if (pkt_len < rx_copybreak &&
 | 
				
			||||||
			    (skb = dev_alloc_skb(pkt_len + 2)) != NULL) {
 | 
								    (skb = netdev_alloc_skb(dev, pkt_len + 2)) != NULL) {
 | 
				
			||||||
				skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
									skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
				
			||||||
				dma_sync_single_for_cpu(&np->pci_dev->dev,
 | 
									dma_sync_single_for_cpu(&np->pci_dev->dev,
 | 
				
			||||||
						le32_to_cpu(desc->frag[0].addr),
 | 
											le32_to_cpu(desc->frag[0].addr),
 | 
				
			||||||
| 
						 | 
					@ -1411,11 +1411,10 @@ static void refill_rx (struct net_device *dev)
 | 
				
			||||||
		struct sk_buff *skb;
 | 
							struct sk_buff *skb;
 | 
				
			||||||
		entry = np->dirty_rx % RX_RING_SIZE;
 | 
							entry = np->dirty_rx % RX_RING_SIZE;
 | 
				
			||||||
		if (np->rx_skbuff[entry] == NULL) {
 | 
							if (np->rx_skbuff[entry] == NULL) {
 | 
				
			||||||
			skb = dev_alloc_skb(np->rx_buf_sz + 2);
 | 
								skb = netdev_alloc_skb(dev, np->rx_buf_sz + 2);
 | 
				
			||||||
			np->rx_skbuff[entry] = skb;
 | 
								np->rx_skbuff[entry] = skb;
 | 
				
			||||||
			if (skb == NULL)
 | 
								if (skb == NULL)
 | 
				
			||||||
				break;		/* Better luck next round. */
 | 
									break;		/* Better luck next round. */
 | 
				
			||||||
			skb->dev = dev;		/* Mark as being used by this device. */
 | 
					 | 
				
			||||||
			skb_reserve(skb, 2);	/* Align IP on 16 byte boundaries */
 | 
								skb_reserve(skb, 2);	/* Align IP on 16 byte boundaries */
 | 
				
			||||||
			np->rx_ring[entry].frag[0].addr = cpu_to_le32(
 | 
								np->rx_ring[entry].frag[0].addr = cpu_to_le32(
 | 
				
			||||||
				dma_map_single(&np->pci_dev->dev, skb->data,
 | 
									dma_map_single(&np->pci_dev->dev, skb->data,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -421,7 +421,7 @@ static int dnet_poll(struct napi_struct *napi, int budget)
 | 
				
			||||||
			printk(KERN_ERR "%s packet receive error %x\n",
 | 
								printk(KERN_ERR "%s packet receive error %x\n",
 | 
				
			||||||
			       __func__, cmd_word);
 | 
								       __func__, cmd_word);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		skb = dev_alloc_skb(pkt_len + 5);
 | 
							skb = netdev_alloc_skb(dev, pkt_len + 5);
 | 
				
			||||||
		if (skb != NULL) {
 | 
							if (skb != NULL) {
 | 
				
			||||||
			/* Align IP on 16 byte boundaries */
 | 
								/* Align IP on 16 byte boundaries */
 | 
				
			||||||
			skb_reserve(skb, 2);
 | 
								skb_reserve(skb, 2);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1070,14 +1070,13 @@ static void allocate_rx_buffers(struct net_device *dev)
 | 
				
			||||||
	while (np->really_rx_count != RX_RING_SIZE) {
 | 
						while (np->really_rx_count != RX_RING_SIZE) {
 | 
				
			||||||
		struct sk_buff *skb;
 | 
							struct sk_buff *skb;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		skb = dev_alloc_skb(np->rx_buf_sz);
 | 
							skb = netdev_alloc_skb(dev, np->rx_buf_sz);
 | 
				
			||||||
		if (skb == NULL)
 | 
							if (skb == NULL)
 | 
				
			||||||
			break;	/* Better luck next round. */
 | 
								break;	/* Better luck next round. */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		while (np->lack_rxbuf->skbuff)
 | 
							while (np->lack_rxbuf->skbuff)
 | 
				
			||||||
			np->lack_rxbuf = np->lack_rxbuf->next_desc_logical;
 | 
								np->lack_rxbuf = np->lack_rxbuf->next_desc_logical;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		skb->dev = dev;	/* Mark as being used by this device. */
 | 
					 | 
				
			||||||
		np->lack_rxbuf->skbuff = skb;
 | 
							np->lack_rxbuf->skbuff = skb;
 | 
				
			||||||
		np->lack_rxbuf->buffer = pci_map_single(np->pci_dev, skb->data,
 | 
							np->lack_rxbuf->buffer = pci_map_single(np->pci_dev, skb->data,
 | 
				
			||||||
			np->rx_buf_sz, PCI_DMA_FROMDEVICE);
 | 
								np->rx_buf_sz, PCI_DMA_FROMDEVICE);
 | 
				
			||||||
| 
						 | 
					@ -1265,7 +1264,7 @@ static void init_ring(struct net_device *dev)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* allocate skb for rx buffers */
 | 
						/* allocate skb for rx buffers */
 | 
				
			||||||
	for (i = 0; i < RX_RING_SIZE; i++) {
 | 
						for (i = 0; i < RX_RING_SIZE; i++) {
 | 
				
			||||||
		struct sk_buff *skb = dev_alloc_skb(np->rx_buf_sz);
 | 
							struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (skb == NULL) {
 | 
							if (skb == NULL) {
 | 
				
			||||||
			np->lack_rxbuf = &np->rx_ring[i];
 | 
								np->lack_rxbuf = &np->rx_ring[i];
 | 
				
			||||||
| 
						 | 
					@ -1274,7 +1273,6 @@ static void init_ring(struct net_device *dev)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		++np->really_rx_count;
 | 
							++np->really_rx_count;
 | 
				
			||||||
		np->rx_ring[i].skbuff = skb;
 | 
							np->rx_ring[i].skbuff = skb;
 | 
				
			||||||
		skb->dev = dev;	/* Mark as being used by this device. */
 | 
					 | 
				
			||||||
		np->rx_ring[i].buffer = pci_map_single(np->pci_dev, skb->data,
 | 
							np->rx_ring[i].buffer = pci_map_single(np->pci_dev, skb->data,
 | 
				
			||||||
			np->rx_buf_sz, PCI_DMA_FROMDEVICE);
 | 
								np->rx_buf_sz, PCI_DMA_FROMDEVICE);
 | 
				
			||||||
		np->rx_ring[i].status = RXOWN;
 | 
							np->rx_ring[i].status = RXOWN;
 | 
				
			||||||
| 
						 | 
					@ -1704,7 +1702,7 @@ static int netdev_rx(struct net_device *dev)
 | 
				
			||||||
			/* Check if the packet is long enough to accept without copying
 | 
								/* Check if the packet is long enough to accept without copying
 | 
				
			||||||
			   to a minimally-sized skbuff. */
 | 
								   to a minimally-sized skbuff. */
 | 
				
			||||||
			if (pkt_len < rx_copybreak &&
 | 
								if (pkt_len < rx_copybreak &&
 | 
				
			||||||
			    (skb = dev_alloc_skb(pkt_len + 2)) != NULL) {
 | 
								    (skb = netdev_alloc_skb(dev, pkt_len + 2)) != NULL) {
 | 
				
			||||||
				skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
									skb_reserve(skb, 2);	/* 16 byte align the IP header */
 | 
				
			||||||
				pci_dma_sync_single_for_cpu(np->pci_dev,
 | 
									pci_dma_sync_single_for_cpu(np->pci_dev,
 | 
				
			||||||
							    np->cur_rx->buffer,
 | 
												    np->cur_rx->buffer,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -711,7 +711,7 @@ fec_enet_rx(struct net_device *ndev)
 | 
				
			||||||
		 * include that when passing upstream as it messes up
 | 
							 * include that when passing upstream as it messes up
 | 
				
			||||||
		 * bridging applications.
 | 
							 * bridging applications.
 | 
				
			||||||
		 */
 | 
							 */
 | 
				
			||||||
		skb = dev_alloc_skb(pkt_len - 4 + NET_IP_ALIGN);
 | 
							skb = netdev_alloc_skb(dev, pkt_len - 4 + NET_IP_ALIGN);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (unlikely(!skb)) {
 | 
							if (unlikely(!skb)) {
 | 
				
			||||||
			printk("%s: Memory squeeze, dropping packet.\n",
 | 
								printk("%s: Memory squeeze, dropping packet.\n",
 | 
				
			||||||
| 
						 | 
					@ -1210,7 +1210,7 @@ static int fec_enet_alloc_buffers(struct net_device *ndev)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	bdp = fep->rx_bd_base;
 | 
						bdp = fep->rx_bd_base;
 | 
				
			||||||
	for (i = 0; i < RX_RING_SIZE; i++) {
 | 
						for (i = 0; i < RX_RING_SIZE; i++) {
 | 
				
			||||||
		skb = dev_alloc_skb(FEC_ENET_RX_FRSIZE);
 | 
							skb = netdev_alloc_skb(dev, FEC_ENET_RX_FRSIZE);
 | 
				
			||||||
		if (!skb) {
 | 
							if (!skb) {
 | 
				
			||||||
			fec_enet_free_buffers(ndev);
 | 
								fec_enet_free_buffers(ndev);
 | 
				
			||||||
			return -ENOMEM;
 | 
								return -ENOMEM;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -160,7 +160,7 @@ static int mpc52xx_fec_alloc_rx_buffers(struct net_device *dev, struct bcom_task
 | 
				
			||||||
	struct sk_buff *skb;
 | 
						struct sk_buff *skb;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	while (!bcom_queue_full(rxtsk)) {
 | 
						while (!bcom_queue_full(rxtsk)) {
 | 
				
			||||||
		skb = dev_alloc_skb(FEC_RX_BUFFER_SIZE);
 | 
							skb = netdev_alloc_skb(dev, FEC_RX_BUFFER_SIZE);
 | 
				
			||||||
		if (!skb)
 | 
							if (!skb)
 | 
				
			||||||
			return -EAGAIN;
 | 
								return -EAGAIN;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -416,7 +416,7 @@ static irqreturn_t mpc52xx_fec_rx_interrupt(int irq, void *dev_id)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* skbs are allocated on open, so now we allocate a new one,
 | 
							/* skbs are allocated on open, so now we allocate a new one,
 | 
				
			||||||
		 * and remove the old (with the packet) */
 | 
							 * and remove the old (with the packet) */
 | 
				
			||||||
		skb = dev_alloc_skb(FEC_RX_BUFFER_SIZE);
 | 
							skb = netdev_alloc_skb(dev, FEC_RX_BUFFER_SIZE);
 | 
				
			||||||
		if (!skb) {
 | 
							if (!skb) {
 | 
				
			||||||
			/* Can't get a new one : reuse the same & drop pkt */
 | 
								/* Can't get a new one : reuse the same & drop pkt */
 | 
				
			||||||
			dev_notice(&dev->dev, "Low memory - dropped packet.\n");
 | 
								dev_notice(&dev->dev, "Low memory - dropped packet.\n");
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -154,7 +154,7 @@ static int fs_enet_rx_napi(struct napi_struct *napi, int budget)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			if (pkt_len <= fpi->rx_copybreak) {
 | 
								if (pkt_len <= fpi->rx_copybreak) {
 | 
				
			||||||
				/* +2 to make IP header L1 cache aligned */
 | 
									/* +2 to make IP header L1 cache aligned */
 | 
				
			||||||
				skbn = dev_alloc_skb(pkt_len + 2);
 | 
									skbn = netdev_alloc_skb(dev, pkt_len + 2);
 | 
				
			||||||
				if (skbn != NULL) {
 | 
									if (skbn != NULL) {
 | 
				
			||||||
					skb_reserve(skbn, 2);	/* align IP header */
 | 
										skb_reserve(skbn, 2);	/* align IP header */
 | 
				
			||||||
					skb_copy_from_linear_data(skb,
 | 
										skb_copy_from_linear_data(skb,
 | 
				
			||||||
| 
						 | 
					@ -165,7 +165,7 @@ static int fs_enet_rx_napi(struct napi_struct *napi, int budget)
 | 
				
			||||||
					skbn = skbt;
 | 
										skbn = skbt;
 | 
				
			||||||
				}
 | 
									}
 | 
				
			||||||
			} else {
 | 
								} else {
 | 
				
			||||||
				skbn = dev_alloc_skb(ENET_RX_FRSIZE);
 | 
									skbn = netdev_alloc_skb(dev, ENET_RX_FRSIZE);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
				if (skbn)
 | 
									if (skbn)
 | 
				
			||||||
					skb_align(skbn, ENET_RX_ALIGN);
 | 
										skb_align(skbn, ENET_RX_ALIGN);
 | 
				
			||||||
| 
						 | 
					@ -286,7 +286,7 @@ static int fs_enet_rx_non_napi(struct net_device *dev)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			if (pkt_len <= fpi->rx_copybreak) {
 | 
								if (pkt_len <= fpi->rx_copybreak) {
 | 
				
			||||||
				/* +2 to make IP header L1 cache aligned */
 | 
									/* +2 to make IP header L1 cache aligned */
 | 
				
			||||||
				skbn = dev_alloc_skb(pkt_len + 2);
 | 
									skbn = netdev_alloc_skb(dev, pkt_len + 2);
 | 
				
			||||||
				if (skbn != NULL) {
 | 
									if (skbn != NULL) {
 | 
				
			||||||
					skb_reserve(skbn, 2);	/* align IP header */
 | 
										skb_reserve(skbn, 2);	/* align IP header */
 | 
				
			||||||
					skb_copy_from_linear_data(skb,
 | 
										skb_copy_from_linear_data(skb,
 | 
				
			||||||
| 
						 | 
					@ -297,7 +297,7 @@ static int fs_enet_rx_non_napi(struct net_device *dev)
 | 
				
			||||||
					skbn = skbt;
 | 
										skbn = skbt;
 | 
				
			||||||
				}
 | 
									}
 | 
				
			||||||
			} else {
 | 
								} else {
 | 
				
			||||||
				skbn = dev_alloc_skb(ENET_RX_FRSIZE);
 | 
									skbn = netdev_alloc_skb(dev, ENET_RX_FRSIZE);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
				if (skbn)
 | 
									if (skbn)
 | 
				
			||||||
					skb_align(skbn, ENET_RX_ALIGN);
 | 
										skb_align(skbn, ENET_RX_ALIGN);
 | 
				
			||||||
| 
						 | 
					@ -504,7 +504,7 @@ void fs_init_bds(struct net_device *dev)
 | 
				
			||||||
	 * Initialize the receive buffer descriptors.
 | 
						 * Initialize the receive buffer descriptors.
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	for (i = 0, bdp = fep->rx_bd_base; i < fep->rx_ring; i++, bdp++) {
 | 
						for (i = 0, bdp = fep->rx_bd_base; i < fep->rx_ring; i++, bdp++) {
 | 
				
			||||||
		skb = dev_alloc_skb(ENET_RX_FRSIZE);
 | 
							skb = netdev_alloc_skb(dev, ENET_RX_FRSIZE);
 | 
				
			||||||
		if (skb == NULL) {
 | 
							if (skb == NULL) {
 | 
				
			||||||
			dev_warn(fep->dev,
 | 
								dev_warn(fep->dev,
 | 
				
			||||||
				 "Memory squeeze, unable to allocate skb\n");
 | 
									 "Memory squeeze, unable to allocate skb\n");
 | 
				
			||||||
| 
						 | 
					@ -592,7 +592,7 @@ static struct sk_buff *tx_skb_align_workaround(struct net_device *dev,
 | 
				
			||||||
	struct fs_enet_private *fep = netdev_priv(dev);
 | 
						struct fs_enet_private *fep = netdev_priv(dev);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Alloc new skb */
 | 
						/* Alloc new skb */
 | 
				
			||||||
	new_skb = dev_alloc_skb(skb->len + 4);
 | 
						new_skb = netdev_alloc_skb(dev, skb->len + 4);
 | 
				
			||||||
	if (!new_skb) {
 | 
						if (!new_skb) {
 | 
				
			||||||
		if (net_ratelimit()) {
 | 
							if (net_ratelimit()) {
 | 
				
			||||||
			dev_warn(fep->dev,
 | 
								dev_warn(fep->dev,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -214,8 +214,9 @@ static struct sk_buff *get_new_skb(struct ucc_geth_private *ugeth,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	skb = __skb_dequeue(&ugeth->rx_recycle);
 | 
						skb = __skb_dequeue(&ugeth->rx_recycle);
 | 
				
			||||||
	if (!skb)
 | 
						if (!skb)
 | 
				
			||||||
		skb = dev_alloc_skb(ugeth->ug_info->uf_info.max_rx_buf_length +
 | 
							skb = netdev_alloc_skb(ugeth->ndev,
 | 
				
			||||||
				    UCC_GETH_RX_DATA_BUF_ALIGNMENT);
 | 
									      ugeth->ug_info->uf_info.max_rx_buf_length +
 | 
				
			||||||
 | 
									      UCC_GETH_RX_DATA_BUF_ALIGNMENT);
 | 
				
			||||||
	if (skb == NULL)
 | 
						if (skb == NULL)
 | 
				
			||||||
		return NULL;
 | 
							return NULL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -227,8 +228,6 @@ static struct sk_buff *get_new_skb(struct ucc_geth_private *ugeth,
 | 
				
			||||||
		    (((unsigned)skb->data) & (UCC_GETH_RX_DATA_BUF_ALIGNMENT -
 | 
							    (((unsigned)skb->data) & (UCC_GETH_RX_DATA_BUF_ALIGNMENT -
 | 
				
			||||||
					      1)));
 | 
										      1)));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	skb->dev = ugeth->ndev;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	out_be32(&((struct qe_bd __iomem *)bd)->buf,
 | 
						out_be32(&((struct qe_bd __iomem *)bd)->buf,
 | 
				
			||||||
		      dma_map_single(ugeth->dev,
 | 
							      dma_map_single(ugeth->dev,
 | 
				
			||||||
				     skb->data,
 | 
									     skb->data,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -757,7 +757,7 @@ net_rx(struct net_device *dev)
 | 
				
			||||||
				dev->stats.rx_errors++;
 | 
									dev->stats.rx_errors++;
 | 
				
			||||||
				break;
 | 
									break;
 | 
				
			||||||
			}
 | 
								}
 | 
				
			||||||
			skb = dev_alloc_skb(pkt_len+3);
 | 
								skb = netdev_alloc_skb(dev, pkt_len + 3);
 | 
				
			||||||
			if (skb == NULL) {
 | 
								if (skb == NULL) {
 | 
				
			||||||
				printk("%s: Memory squeeze, dropping packet (len %d).\n",
 | 
									printk("%s: Memory squeeze, dropping packet (len %d).\n",
 | 
				
			||||||
					   dev->name, pkt_len);
 | 
										   dev->name, pkt_len);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue