mv643xx_eth: move rx_return_buff() into its only caller
authorLennert Buytenhek <buytenh@wantstofly.org>
Sun, 1 Jun 2008 08:07:49 +0000 (10:07 +0200)
committerLennert Buytenhek <buytenh@wantstofly.org>
Thu, 12 Jun 2008 06:40:32 +0000 (08:40 +0200)
rx_return_buff() is also a remnant of the HAL layering that the
original mv643xx_eth driver used.  Moving it into its caller kills
the last reference to FUNC_RET_STATUS/pkt_info.

Signed-off-by: Lennert Buytenhek <buytenh@marvell.com>
Acked-by: Dale Farnsworth <dale@farnsworth.org>
drivers/net/mv643xx_eth.c

index 58c1c1b0aa547f5ae382eb651fa2429ffb4a5f62..0e4babbe37ae3a5192b14c7d25cea59df86b3523 100644 (file)
@@ -443,74 +443,50 @@ static unsigned int mv643xx_eth_port_disable_tx(struct mv643xx_eth_private *mp)
 /* rx ***********************************************************************/
 static void mv643xx_eth_free_completed_tx_descs(struct net_device *dev);
 
-static FUNC_RET_STATUS rx_return_buff(struct mv643xx_eth_private *mp,
-                                               struct pkt_info *pkt_info)
+static void mv643xx_eth_rx_refill_descs(struct net_device *dev)
 {
-       int used_rx_desc;       /* Where to return Rx resource */
-       volatile struct rx_desc *rx_desc;
+       struct mv643xx_eth_private *mp = netdev_priv(dev);
        unsigned long flags;
 
        spin_lock_irqsave(&mp->lock, flags);
 
-       /* Get 'used' Rx descriptor */
-       used_rx_desc = mp->rx_used_desc;
-       rx_desc = &mp->rx_desc_area[used_rx_desc];
-
-       rx_desc->buf_ptr = pkt_info->buf_ptr;
-       rx_desc->buf_size = pkt_info->byte_cnt;
-       mp->rx_skb[used_rx_desc] = pkt_info->return_info;
-
-       /* Flush the write pipe */
-
-       /* Return the descriptor to DMA ownership */
-       wmb();
-       rx_desc->cmd_sts = BUFFER_OWNED_BY_DMA | RX_ENABLE_INTERRUPT;
-       wmb();
-
-       /* Move the used descriptor pointer to the next descriptor */
-       mp->rx_used_desc = (used_rx_desc + 1) % mp->rx_ring_size;
-
-       spin_unlock_irqrestore(&mp->lock, flags);
-
-       return ETH_OK;
-}
-
-static void mv643xx_eth_rx_refill_descs(struct net_device *dev)
-{
-       struct mv643xx_eth_private *mp = netdev_priv(dev);
-       struct pkt_info pkt_info;
-       struct sk_buff *skb;
-       int unaligned;
-
        while (mp->rx_desc_count < mp->rx_ring_size) {
+               struct sk_buff *skb;
+               int unaligned;
+               int rx;
+
                skb = dev_alloc_skb(ETH_RX_SKB_SIZE + dma_get_cache_alignment());
-               if (!skb)
+               if (skb == NULL)
                        break;
-               mp->rx_desc_count++;
+
                unaligned = (u32)skb->data & (dma_get_cache_alignment() - 1);
                if (unaligned)
                        skb_reserve(skb, dma_get_cache_alignment() - unaligned);
-               pkt_info.cmd_sts = RX_ENABLE_INTERRUPT;
-               pkt_info.byte_cnt = ETH_RX_SKB_SIZE;
-               pkt_info.buf_ptr = dma_map_single(NULL, skb->data,
-                                       ETH_RX_SKB_SIZE, DMA_FROM_DEVICE);
-               pkt_info.return_info = skb;
-               if (rx_return_buff(mp, &pkt_info) != ETH_OK) {
-                       printk(KERN_ERR
-                               "%s: Error allocating RX Ring\n", dev->name);
-                       break;
-               }
+
+               mp->rx_desc_count++;
+               rx = mp->rx_used_desc;
+               mp->rx_used_desc = (rx + 1) % mp->rx_ring_size;
+
+               mp->rx_desc_area[rx].buf_ptr = dma_map_single(NULL,
+                                                       skb->data,
+                                                       ETH_RX_SKB_SIZE,
+                                                       DMA_FROM_DEVICE);
+               mp->rx_desc_area[rx].buf_size = ETH_RX_SKB_SIZE;
+               mp->rx_skb[rx] = skb;
+               wmb();
+               mp->rx_desc_area[rx].cmd_sts = BUFFER_OWNED_BY_DMA |
+                                               RX_ENABLE_INTERRUPT;
+               wmb();
+
                skb_reserve(skb, ETH_HW_IP_ALIGN);
        }
-       /*
-        * If RX ring is empty of SKB, set a timer to try allocating
-        * again at a later time.
-        */
+
        if (mp->rx_desc_count == 0) {
-               printk(KERN_INFO "%s: Rx ring is empty\n", dev->name);
-               mp->timeout.expires = jiffies + (HZ / 10);      /* 100 mSec */
+               mp->timeout.expires = jiffies + (HZ / 10);
                add_timer(&mp->timeout);
        }
+
+       spin_unlock_irqrestore(&mp->lock, flags);
 }
 
 static inline void mv643xx_eth_rx_refill_descs_timer_wrapper(unsigned long data)
This page took 0.028809 seconds and 5 git commands to generate.