[PATCH] forcedeth: Jumbo Frame Support

This is a multi-part message in MIME format.
Signed-off-by: Jeff Garzik <jgarzik@pobox.com>
This commit is contained in:
Manfred Spraul 2005-07-31 18:20:30 +02:00 committed by Jeff Garzik
parent 0e920bfb03
commit d81c0983de

View file

@ -85,6 +85,7 @@
* 0.33: 16 May 2005: Support for MCP51 added. * 0.33: 16 May 2005: Support for MCP51 added.
* 0.34: 18 Jun 2005: Add DEV_NEED_LINKTIMER to all nForce nics. * 0.34: 18 Jun 2005: Add DEV_NEED_LINKTIMER to all nForce nics.
* 0.35: 26 Jun 2005: Support for MCP55 added. * 0.35: 26 Jun 2005: Support for MCP55 added.
* 0.36: 28 Jul 2005: Add jumbo frame support.
* *
* Known bugs: * Known bugs:
* We suspect that on some hardware no TX done interrupts are generated. * We suspect that on some hardware no TX done interrupts are generated.
@ -96,7 +97,7 @@
* DEV_NEED_TIMERIRQ will not harm you on sane hardware, only generating a few * DEV_NEED_TIMERIRQ will not harm you on sane hardware, only generating a few
* superfluous timer interrupts from the nic. * superfluous timer interrupts from the nic.
*/ */
#define FORCEDETH_VERSION "0.35" #define FORCEDETH_VERSION "0.36"
#define DRV_NAME "forcedeth" #define DRV_NAME "forcedeth"
#include <linux/module.h> #include <linux/module.h>
@ -379,9 +380,13 @@ struct ring_desc {
#define TX_LIMIT_START 62 #define TX_LIMIT_START 62
/* rx/tx mac addr + type + vlan + align + slack*/ /* rx/tx mac addr + type + vlan + align + slack*/
#define RX_NIC_BUFSIZE (ETH_DATA_LEN + 64) #define NV_RX_HEADERS (64)
/* even more slack */ /* even more slack. */
#define RX_ALLOC_BUFSIZE (ETH_DATA_LEN + 128) #define NV_RX_ALLOC_PAD (64)
/* maximum mtu size */
#define NV_PKTLIMIT_1 ETH_DATA_LEN /* hard limit not known */
#define NV_PKTLIMIT_2 9100 /* Actual limit according to NVidia: 9202 */
#define OOM_REFILL (1+HZ/20) #define OOM_REFILL (1+HZ/20)
#define POLL_WAIT (1+HZ/100) #define POLL_WAIT (1+HZ/100)
@ -473,6 +478,7 @@ struct fe_priv {
struct sk_buff *rx_skbuff[RX_RING]; struct sk_buff *rx_skbuff[RX_RING];
dma_addr_t rx_dma[RX_RING]; dma_addr_t rx_dma[RX_RING];
unsigned int rx_buf_sz; unsigned int rx_buf_sz;
unsigned int pkt_limit;
struct timer_list oom_kick; struct timer_list oom_kick;
struct timer_list nic_poll; struct timer_list nic_poll;
@ -792,7 +798,7 @@ static int nv_alloc_rx(struct net_device *dev)
nr = refill_rx % RX_RING; nr = refill_rx % RX_RING;
if (np->rx_skbuff[nr] == NULL) { if (np->rx_skbuff[nr] == NULL) {
skb = dev_alloc_skb(RX_ALLOC_BUFSIZE); skb = dev_alloc_skb(np->rx_buf_sz + NV_RX_ALLOC_PAD);
if (!skb) if (!skb)
break; break;
@ -805,7 +811,7 @@ static int nv_alloc_rx(struct net_device *dev)
PCI_DMA_FROMDEVICE); PCI_DMA_FROMDEVICE);
np->rx_ring[nr].PacketBuffer = cpu_to_le32(np->rx_dma[nr]); np->rx_ring[nr].PacketBuffer = cpu_to_le32(np->rx_dma[nr]);
wmb(); wmb();
np->rx_ring[nr].FlagLen = cpu_to_le32(RX_NIC_BUFSIZE | NV_RX_AVAIL); np->rx_ring[nr].FlagLen = cpu_to_le32(np->rx_buf_sz | NV_RX_AVAIL);
dprintk(KERN_DEBUG "%s: nv_alloc_rx: Packet %d marked as Available\n", dprintk(KERN_DEBUG "%s: nv_alloc_rx: Packet %d marked as Available\n",
dev->name, refill_rx); dev->name, refill_rx);
refill_rx++; refill_rx++;
@ -831,7 +837,18 @@ static void nv_do_rx_refill(unsigned long data)
enable_irq(dev->irq); enable_irq(dev->irq);
} }
static int nv_init_ring(struct net_device *dev) static void nv_init_rx(struct net_device *dev)
{
struct fe_priv *np = get_nvpriv(dev);
int i;
np->cur_rx = RX_RING;
np->refill_rx = 0;
for (i = 0; i < RX_RING; i++)
np->rx_ring[i].FlagLen = 0;
}
static void nv_init_tx(struct net_device *dev)
{ {
struct fe_priv *np = get_nvpriv(dev); struct fe_priv *np = get_nvpriv(dev);
int i; int i;
@ -839,11 +856,12 @@ static int nv_init_ring(struct net_device *dev)
np->next_tx = np->nic_tx = 0; np->next_tx = np->nic_tx = 0;
for (i = 0; i < TX_RING; i++) for (i = 0; i < TX_RING; i++)
np->tx_ring[i].FlagLen = 0; np->tx_ring[i].FlagLen = 0;
}
np->cur_rx = RX_RING; static int nv_init_ring(struct net_device *dev)
np->refill_rx = 0; {
for (i = 0; i < RX_RING; i++) nv_init_tx(dev);
np->rx_ring[i].FlagLen = 0; nv_init_rx(dev);
return nv_alloc_rx(dev); return nv_alloc_rx(dev);
} }
@ -1207,15 +1225,82 @@ next_pkt:
} }
} }
static void set_bufsize(struct net_device *dev)
{
struct fe_priv *np = netdev_priv(dev);
if (dev->mtu <= ETH_DATA_LEN)
np->rx_buf_sz = ETH_DATA_LEN + NV_RX_HEADERS;
else
np->rx_buf_sz = dev->mtu + NV_RX_HEADERS;
}
/* /*
* nv_change_mtu: dev->change_mtu function * nv_change_mtu: dev->change_mtu function
* Called with dev_base_lock held for read. * Called with dev_base_lock held for read.
*/ */
static int nv_change_mtu(struct net_device *dev, int new_mtu) static int nv_change_mtu(struct net_device *dev, int new_mtu)
{ {
if (new_mtu > ETH_DATA_LEN) struct fe_priv *np = get_nvpriv(dev);
int old_mtu;
if (new_mtu < 64 || new_mtu > np->pkt_limit)
return -EINVAL; return -EINVAL;
old_mtu = dev->mtu;
dev->mtu = new_mtu; dev->mtu = new_mtu;
/* return early if the buffer sizes will not change */
if (old_mtu <= ETH_DATA_LEN && new_mtu <= ETH_DATA_LEN)
return 0;
if (old_mtu == new_mtu)
return 0;
/* synchronized against open : rtnl_lock() held by caller */
if (netif_running(dev)) {
u8 *base = get_hwbase(dev);
/*
* It seems that the nic preloads valid ring entries into an
* internal buffer. The procedure for flushing everything is
* guessed, there is probably a simpler approach.
* Changing the MTU is a rare event, it shouldn't matter.
*/
disable_irq(dev->irq);
spin_lock_bh(&dev->xmit_lock);
spin_lock(&np->lock);
/* stop engines */
nv_stop_rx(dev);
nv_stop_tx(dev);
nv_txrx_reset(dev);
/* drain rx queue */
nv_drain_rx(dev);
nv_drain_tx(dev);
/* reinit driver view of the rx queue */
nv_init_rx(dev);
nv_init_tx(dev);
/* alloc new rx buffers */
set_bufsize(dev);
if (nv_alloc_rx(dev)) {
if (!np->in_shutdown)
mod_timer(&np->oom_kick, jiffies + OOM_REFILL);
}
/* reinit nic view of the rx queue */
writel(np->rx_buf_sz, base + NvRegOffloadConfig);
writel((u32) np->ring_addr, base + NvRegRxRingPhysAddr);
writel((u32) (np->ring_addr + RX_RING*sizeof(struct ring_desc)), base + NvRegTxRingPhysAddr);
writel( ((RX_RING-1) << NVREG_RINGSZ_RXSHIFT) + ((TX_RING-1) << NVREG_RINGSZ_TXSHIFT),
base + NvRegRingSizes);
pci_push(base);
writel(NVREG_TXRXCTL_KICK|np->desc_ver, get_hwbase(dev) + NvRegTxRxControl);
pci_push(base);
/* restart rx engine */
nv_start_rx(dev);
nv_start_tx(dev);
spin_unlock(&np->lock);
spin_unlock_bh(&dev->xmit_lock);
enable_irq(dev->irq);
}
return 0; return 0;
} }
@ -1792,6 +1877,7 @@ static int nv_open(struct net_device *dev)
writel(0, base + NvRegAdapterControl); writel(0, base + NvRegAdapterControl);
/* 2) initialize descriptor rings */ /* 2) initialize descriptor rings */
set_bufsize(dev);
oom = nv_init_ring(dev); oom = nv_init_ring(dev);
writel(0, base + NvRegLinkSpeed); writel(0, base + NvRegLinkSpeed);
@ -1837,7 +1923,7 @@ static int nv_open(struct net_device *dev)
writel(NVREG_MISC1_FORCE | NVREG_MISC1_HD, base + NvRegMisc1); writel(NVREG_MISC1_FORCE | NVREG_MISC1_HD, base + NvRegMisc1);
writel(readl(base + NvRegTransmitterStatus), base + NvRegTransmitterStatus); writel(readl(base + NvRegTransmitterStatus), base + NvRegTransmitterStatus);
writel(NVREG_PFF_ALWAYS, base + NvRegPacketFilterFlags); writel(NVREG_PFF_ALWAYS, base + NvRegPacketFilterFlags);
writel(NVREG_OFFLOAD_NORMAL, base + NvRegOffloadConfig); writel(np->rx_buf_sz, base + NvRegOffloadConfig);
writel(readl(base + NvRegReceiverStatus), base + NvRegReceiverStatus); writel(readl(base + NvRegReceiverStatus), base + NvRegReceiverStatus);
get_random_bytes(&i, sizeof(i)); get_random_bytes(&i, sizeof(i));
@ -2007,13 +2093,16 @@ static int __devinit nv_probe(struct pci_dev *pci_dev, const struct pci_device_i
/* handle different descriptor versions */ /* handle different descriptor versions */
if (pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_1 || if (pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_1 ||
pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_2 || pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_2 ||
pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_3 || pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_3 ||
pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_12 || pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_12 ||
pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_13) pci_dev->device == PCI_DEVICE_ID_NVIDIA_NVENET_13) {
np->desc_ver = DESC_VER_1; np->desc_ver = DESC_VER_1;
else np->pkt_limit = NV_PKTLIMIT_1;
} else {
np->desc_ver = DESC_VER_2; np->desc_ver = DESC_VER_2;
np->pkt_limit = NV_PKTLIMIT_2;
}
err = -ENOMEM; err = -ENOMEM;
np->base = ioremap(addr, NV_PCI_REGSZ); np->base = ioremap(addr, NV_PCI_REGSZ);