ramips_eth: coding style cleanup

SVN-Revision: 19414
owl
Gabor Juhos 2010-01-30 15:25:55 +00:00
parent a4e85bbceb
commit 67cd819da6
1 changed files with 66 additions and 54 deletions

View File

@ -83,12 +83,12 @@ ramips_alloc_dma(struct net_device *dev)
/* setup tx ring */
priv->tx = dma_alloc_coherent(NULL,
NUM_TX_DESC * sizeof(struct ramips_tx_dma), &priv->phy_tx, GFP_ATOMIC);
NUM_TX_DESC * sizeof(struct ramips_tx_dma),
&priv->phy_tx, GFP_ATOMIC);
if (!priv->tx)
goto err_cleanup;
for(i = 0; i < NUM_TX_DESC; i++)
{
for (i = 0; i < NUM_TX_DESC; i++) {
memset(&priv->tx[i], 0, sizeof(struct ramips_tx_dma));
priv->tx[i].txd2 |= TX_DMA_LSO | TX_DMA_DONE;
priv->tx[i].txd4 &= (TX_DMA_QN_MASK | TX_DMA_PN_MASK);
@ -97,21 +97,22 @@ ramips_alloc_dma(struct net_device *dev)
/* setup rx ring */
priv->rx = dma_alloc_coherent(NULL,
NUM_RX_DESC * sizeof(struct ramips_rx_dma), &priv->phy_rx, GFP_ATOMIC);
NUM_RX_DESC * sizeof(struct ramips_rx_dma),
&priv->phy_rx, GFP_ATOMIC);
if (!priv->rx)
goto err_cleanup;
memset(priv->rx, 0, sizeof(struct ramips_rx_dma) * NUM_RX_DESC);
for(i = 0; i < NUM_RX_DESC; i++)
{
for (i = 0; i < NUM_RX_DESC; i++) {
struct sk_buff *new_skb = dev_alloc_skb(MAX_RX_LENGTH + 2);
if (!new_skb)
goto err_cleanup;
skb_reserve(new_skb, 2);
priv->rx[i].rxd1 =
dma_map_single(NULL, skb_put(new_skb, 2), MAX_RX_LENGTH + 2,
priv->rx[i].rxd1 = dma_map_single(NULL,
skb_put(new_skb, 2),
MAX_RX_LENGTH + 2,
DMA_FROM_DEVICE);
priv->rx[i].rxd2 |= RX_DMA_LSO;
priv->rx_skb[i] = new_skb;
@ -149,19 +150,18 @@ ramips_eth_hard_start_xmit(struct sk_buff* skb, struct net_device *dev)
unsigned int mapped_addr;
unsigned long flags;
if(priv->plat->min_pkt_len)
{
if(skb->len < priv->plat->min_pkt_len)
{
if(skb_padto(skb, priv->plat->min_pkt_len))
{
printk(KERN_ERR "ramips_eth: skb_padto failed\n");
if (priv->plat->min_pkt_len) {
if (skb->len < priv->plat->min_pkt_len) {
if (skb_padto(skb, priv->plat->min_pkt_len)) {
printk(KERN_ERR
"ramips_eth: skb_padto failed\n");
kfree_skb(skb);
return 0;
}
skb_put(skb, priv->plat->min_pkt_len - skb->len);
}
}
dev->trans_start = jiffies;
mapped_addr = (unsigned int) dma_map_single(NULL, skb->data, skb->len,
DMA_TO_DEVICE);
@ -172,9 +172,12 @@ ramips_eth_hard_start_xmit(struct sk_buff* skb, struct net_device *dev)
tx_next = 0;
else
tx_next = tx + 1;
if ((priv->tx_skb[tx]) || (priv->tx_skb[tx_next]) ||
!(priv->tx[tx].txd2 & TX_DMA_DONE) || !(priv->tx[tx_next].txd2 & TX_DMA_DONE))
!(priv->tx[tx].txd2 & TX_DMA_DONE) ||
!(priv->tx[tx_next].txd2 & TX_DMA_DONE))
goto out;
priv->tx[tx].txd1 = mapped_addr;
priv->tx[tx].txd2 &= ~(TX_DMA_PLEN0_MASK | TX_DMA_DONE);
priv->tx[tx].txd2 |= TX_DMA_PLEN0(skb->len);
@ -185,6 +188,7 @@ ramips_eth_hard_start_xmit(struct sk_buff* skb, struct net_device *dev)
ramips_fe_wr((tx + 1) % NUM_TX_DESC, RAMIPS_TX_CTX_IDX0);
spin_unlock_irqrestore(&priv->page_lock, flags);
return NETDEV_TX_OK;
out:
spin_unlock_irqrestore(&priv->page_lock, flags);
dev->stats.tx_dropped++;
@ -200,8 +204,7 @@ ramips_eth_rx_hw(unsigned long ptr)
int rx;
int max_rx = 16;
while(max_rx)
{
while (max_rx) {
struct sk_buff *rx_skb, *new_skb;
rx = (ramips_fe_rr(RAMIPS_RX_CALC_IDX0) + 1) % NUM_RX_DESC;
@ -222,13 +225,15 @@ ramips_eth_rx_hw(unsigned long ptr)
priv->rx_skb[rx] = new_skb;
BUG_ON(!new_skb);
skb_reserve(new_skb, 2);
priv->rx[rx].rxd1 =
dma_map_single(NULL, new_skb->data, MAX_RX_LENGTH + 2,
priv->rx[rx].rxd1 = dma_map_single(NULL,
new_skb->data,
MAX_RX_LENGTH + 2,
DMA_FROM_DEVICE);
priv->rx[rx].rxd2 &= ~RX_DMA_DONE;
wmb();
ramips_fe_wr(rx, RAMIPS_RX_CALC_IDX0);
}
if (max_rx == 0)
tasklet_schedule(&priv->rx_tasklet);
else
@ -243,14 +248,14 @@ ramips_eth_tx_housekeeping(unsigned long ptr)
struct raeth_priv *priv = netdev_priv(dev);
while ((priv->tx[priv->skb_free_idx].txd2 & TX_DMA_DONE) &&
(priv->tx_skb[priv->skb_free_idx]))
{
(priv->tx_skb[priv->skb_free_idx])) {
dev_kfree_skb_irq((struct sk_buff *) priv->tx_skb[priv->skb_free_idx]);
priv->tx_skb[priv->skb_free_idx] = 0;
priv->skb_free_idx++;
if (priv->skb_free_idx >= NUM_TX_DESC)
priv->skb_free_idx = 0;
}
ramips_fe_wr(ramips_fe_rr(RAMIPS_FE_INT_ENABLE) | RAMIPS_TX_DLY_INT,
RAMIPS_FE_INT_ENABLE);
}
@ -262,6 +267,7 @@ ramips_eth_set_mac_addr(struct net_device *dev, void *priv)
if (netif_running(dev))
return -EBUSY;
memcpy(dev->dev_addr, ((struct sockaddr*)priv)->sa_data, dev->addr_len);
ramips_fe_wr((mac[0] << 8) | mac[1], RAMIPS_GDMA1_MAC_ADRH);
ramips_fe_wr(RAMIPS_GDMA1_MAC_ADRL,
@ -285,14 +291,15 @@ ramips_eth_irq(int irq, void *dev)
ramips_fe_wr(0xFFFFFFFF, RAMIPS_FE_INT_STATUS);
if(fe_int & RAMIPS_RX_DLY_INT)
{
if (fe_int & RAMIPS_RX_DLY_INT) {
ramips_fe_wr(ramips_fe_rr(RAMIPS_FE_INT_ENABLE) & ~(RAMIPS_RX_DLY_INT),
RAMIPS_FE_INT_ENABLE);
tasklet_schedule(&priv->rx_tasklet);
}
if (fe_int & RAMIPS_TX_DLY_INT)
ramips_eth_tx_housekeeping((unsigned long)dev);
return IRQ_HANDLED;
}
@ -320,9 +327,11 @@ ramips_eth_open(struct net_device *dev)
~(RAMIPS_US_CYC_CNT_MASK << RAMIPS_US_CYC_CNT_SHIFT)) |
((rt305x_sys_freq / RAMIPS_US_CYC_CNT_DIVISOR) << RAMIPS_US_CYC_CNT_SHIFT),
RAMIPS_FE_GLO_CFG);
tasklet_init(&priv->tx_housekeeping_tasklet, ramips_eth_tx_housekeeping,
(unsigned long)dev);
tasklet_init(&priv->rx_tasklet, ramips_eth_rx_hw, (unsigned long)dev);
ramips_fe_wr(RAMIPS_DELAY_INIT, RAMIPS_DLY_INT_CFG);
ramips_fe_wr(RAMIPS_TX_DLY_INT | RAMIPS_RX_DLY_INT, RAMIPS_FE_INT_ENABLE);
ramips_fe_wr(ramips_fe_rr(RAMIPS_GDMA1_FWD_CFG) &
@ -334,6 +343,7 @@ ramips_eth_open(struct net_device *dev)
ramips_fe_wr(RAMIPS_PSE_FQFC_CFG_INIT, RAMIPS_PSE_FQ_CFG);
ramips_fe_wr(1, RAMIPS_FE_RST_GL);
ramips_fe_wr(0, RAMIPS_FE_RST_GL);
netif_start_queue(dev);
return 0;
@ -379,6 +389,7 @@ ramips_eth_probe(struct net_device *dev)
dev->tx_timeout = ramips_eth_timeout;
dev->watchdog_timeo = TX_TIMEOUT;
spin_lock_init(&priv->page_lock);
return 0;
}
@ -422,7 +433,8 @@ ramips_eth_plat_probe(struct platform_device *plat)
ramips_dev->addr_len = ETH_ALEN;
ramips_dev->base_addr = (unsigned long)ramips_fe_base;
ramips_dev->init = ramips_eth_probe;
priv = (struct raeth_priv*)netdev_priv(ramips_dev);
priv = netdev_priv(ramips_dev);
priv->plat = data;
err = register_netdev(ramips_dev);