drivers/net: Remove casts of void *

Unnecessary casts of void * clutter the code.

These are the remainder casts after several specific
patches to remove netdev_priv and dev_priv.

Done via coccinelle script (and a little editing):

$ cat cast_void_pointer.cocci
@@
type T;
T *pt;
void *pv;
@@

- pt = (T *)pv;
+ pt = pv;

Signed-off-by: Joe Perches <joe@perches.com>
Acked-by: Sjur Brændeland <sjur.brandeland@stericsson.com>
Acked-By: Chris Snook <chris.snook@gmail.com>
Acked-by: Jon Mason <jdmason@kudzu.us>
Acked-by: Geert Uytterhoeven <geert@linux-m68k.org>
Acked-by: David Dillow <dave@thedillows.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
Joe Perches 2011-06-16 19:08:06 +00:00 committed by David S. Miller
parent dadbe85ac4
commit 43d620c829
41 changed files with 124 additions and 147 deletions

View file

@ -421,16 +421,11 @@ static int lance_tx (struct net_device *dev)
static irqreturn_t lance_interrupt (int irq, void *dev_id) static irqreturn_t lance_interrupt (int irq, void *dev_id)
{ {
struct net_device *dev; struct net_device *dev = dev_id;
struct lance_private *lp; struct lance_private *lp = netdev_priv(dev);
volatile struct lance_regs *ll; volatile struct lance_regs *ll = lp->ll;
int csr0; int csr0;
dev = (struct net_device *) dev_id;
lp = netdev_priv(dev);
ll = lp->ll;
ll->rap = LE_CSR0; /* LANCE Controller Status */ ll->rap = LE_CSR0; /* LANCE Controller Status */
csr0 = ll->rdp; csr0 = ll->rdp;

View file

@ -652,9 +652,9 @@ static int do_write(struct net_device *dev, void *cbuf, int cbuflen,
int ret; int ret;
if(i) { if(i) {
qels[i].cbuf = (unsigned char *) cbuf; qels[i].cbuf = cbuf;
qels[i].cbuflen = cbuflen; qels[i].cbuflen = cbuflen;
qels[i].dbuf = (unsigned char *) dbuf; qels[i].dbuf = dbuf;
qels[i].dbuflen = dbuflen; qels[i].dbuflen = dbuflen;
qels[i].QWrite = 1; qels[i].QWrite = 1;
qels[i].mailbox = i; /* this should be initted rather */ qels[i].mailbox = i; /* this should be initted rather */
@ -676,9 +676,9 @@ static int do_read(struct net_device *dev, void *cbuf, int cbuflen,
int ret; int ret;
if(i) { if(i) {
qels[i].cbuf = (unsigned char *) cbuf; qels[i].cbuf = cbuf;
qels[i].cbuflen = cbuflen; qels[i].cbuflen = cbuflen;
qels[i].dbuf = (unsigned char *) dbuf; qels[i].dbuf = dbuf;
qels[i].dbuflen = dbuflen; qels[i].dbuflen = dbuflen;
qels[i].QWrite = 0; qels[i].QWrite = 0;
qels[i].mailbox = i; /* this should be initted rather */ qels[i].mailbox = i; /* this should be initted rather */

View file

@ -800,8 +800,7 @@ static int atl1e_setup_ring_resources(struct atl1e_adapter *adapter)
/* Init TPD Ring */ /* Init TPD Ring */
tx_ring->dma = roundup(adapter->ring_dma, 8); tx_ring->dma = roundup(adapter->ring_dma, 8);
offset = tx_ring->dma - adapter->ring_dma; offset = tx_ring->dma - adapter->ring_dma;
tx_ring->desc = (struct atl1e_tpd_desc *) tx_ring->desc = adapter->ring_vir_addr + offset;
(adapter->ring_vir_addr + offset);
size = sizeof(struct atl1e_tx_buffer) * (tx_ring->count); size = sizeof(struct atl1e_tx_buffer) * (tx_ring->count);
tx_ring->tx_buffer = kzalloc(size, GFP_KERNEL); tx_ring->tx_buffer = kzalloc(size, GFP_KERNEL);
if (tx_ring->tx_buffer == NULL) { if (tx_ring->tx_buffer == NULL) {
@ -827,7 +826,7 @@ static int atl1e_setup_ring_resources(struct atl1e_adapter *adapter)
/* Init CMB dma address */ /* Init CMB dma address */
tx_ring->cmb_dma = adapter->ring_dma + offset; tx_ring->cmb_dma = adapter->ring_dma + offset;
tx_ring->cmb = (u32 *)(adapter->ring_vir_addr + offset); tx_ring->cmb = adapter->ring_vir_addr + offset;
offset += sizeof(u32); offset += sizeof(u32);
for (i = 0; i < adapter->num_rx_queues; i++) { for (i = 0; i < adapter->num_rx_queues; i++) {

View file

@ -311,8 +311,7 @@ static s32 atl2_setup_ring_resources(struct atl2_adapter *adapter)
adapter->txd_dma = adapter->ring_dma ; adapter->txd_dma = adapter->ring_dma ;
offset = (adapter->txd_dma & 0x7) ? (8 - (adapter->txd_dma & 0x7)) : 0; offset = (adapter->txd_dma & 0x7) ? (8 - (adapter->txd_dma & 0x7)) : 0;
adapter->txd_dma += offset; adapter->txd_dma += offset;
adapter->txd_ring = (struct tx_pkt_header *) (adapter->ring_vir_addr + adapter->txd_ring = adapter->ring_vir_addr + offset;
offset);
/* Init TXS Ring */ /* Init TXS Ring */
adapter->txs_dma = adapter->txd_dma + adapter->txd_ring_size; adapter->txs_dma = adapter->txd_dma + adapter->txd_ring_size;

View file

@ -2334,8 +2334,7 @@ int be_cmd_get_cntl_attributes(struct be_adapter *adapter)
status = be_mbox_notify_wait(adapter); status = be_mbox_notify_wait(adapter);
if (!status) { if (!status) {
attribs = (struct mgmt_controller_attrib *)( attribs_cmd.va + attribs = attribs_cmd.va + sizeof(struct be_cmd_resp_hdr);
sizeof(struct be_cmd_resp_hdr));
adapter->hba_port_num = attribs->hba_attribs.phy_port; adapter->hba_port_num = attribs->hba_attribs.phy_port;
} }

View file

@ -408,7 +408,7 @@ static int be_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
} }
status = be_cmd_get_phy_info(adapter, &phy_cmd); status = be_cmd_get_phy_info(adapter, &phy_cmd);
if (!status) { if (!status) {
resp = (struct be_cmd_resp_get_phy_info *) phy_cmd.va; resp = phy_cmd.va;
intf_type = le16_to_cpu(resp->interface_type); intf_type = le16_to_cpu(resp->interface_type);
switch (intf_type) { switch (intf_type) {
@ -712,7 +712,7 @@ be_read_eeprom(struct net_device *netdev, struct ethtool_eeprom *eeprom,
status = be_cmd_get_seeprom_data(adapter, &eeprom_cmd); status = be_cmd_get_seeprom_data(adapter, &eeprom_cmd);
if (!status) { if (!status) {
resp = (struct be_cmd_resp_seeprom_read *) eeprom_cmd.va; resp = eeprom_cmd.va;
memcpy(data, resp->seeprom_data + eeprom->offset, eeprom->len); memcpy(data, resp->seeprom_data + eeprom->offset, eeprom->len);
} }
dma_free_coherent(&adapter->pdev->dev, eeprom_cmd.size, eeprom_cmd.va, dma_free_coherent(&adapter->pdev->dev, eeprom_cmd.size, eeprom_cmd.va,

View file

@ -236,7 +236,7 @@ static void
bfa_cee_hbfail(void *arg) bfa_cee_hbfail(void *arg)
{ {
struct bfa_cee *cee; struct bfa_cee *cee;
cee = (struct bfa_cee *) arg; cee = arg;
if (cee->get_attr_pending == true) { if (cee->get_attr_pending == true) {
cee->get_attr_status = BFA_STATUS_FAILED; cee->get_attr_status = BFA_STATUS_FAILED;

View file

@ -74,7 +74,7 @@ typedef struct mac { u8 mac[MAC_ADDRLEN]; } mac_t;
bfa_q_next(_q) = bfa_q_next(*((struct list_head **) _qe)); \ bfa_q_next(_q) = bfa_q_next(*((struct list_head **) _qe)); \
bfa_q_qe_init(*((struct list_head **) _qe)); \ bfa_q_qe_init(*((struct list_head **) _qe)); \
} else { \ } else { \
*((struct list_head **) (_qe)) = (struct list_head *) NULL; \ *((struct list_head **)(_qe)) = NULL; \
} \ } \
} }

View file

@ -134,7 +134,7 @@ int caif_shmdrv_rx_cb(u32 mbx_msg, void *priv)
u32 avail_emptybuff = 0; u32 avail_emptybuff = 0;
unsigned long flags = 0; unsigned long flags = 0;
pshm_drv = (struct shmdrv_layer *)priv; pshm_drv = priv;
/* Check for received buffers. */ /* Check for received buffers. */
if (mbx_msg & SHM_FULL_MASK) { if (mbx_msg & SHM_FULL_MASK) {

View file

@ -4318,7 +4318,7 @@ static void cnic_init_bnx2_tx_ring(struct cnic_dev *dev)
val = BNX2_L2CTX_CMD_TYPE_TYPE_L2 | (8 << 16); val = BNX2_L2CTX_CMD_TYPE_TYPE_L2 | (8 << 16);
cnic_ctx_wr(dev, cid_addr, offset1, val); cnic_ctx_wr(dev, cid_addr, offset1, val);
txbd = (struct tx_bd *) udev->l2_ring; txbd = udev->l2_ring;
buf_map = udev->l2_buf_map; buf_map = udev->l2_buf_map;
for (i = 0; i < MAX_TX_DESC_CNT; i++, txbd++) { for (i = 0; i < MAX_TX_DESC_CNT; i++, txbd++) {
@ -4377,7 +4377,7 @@ static void cnic_init_bnx2_rx_ring(struct cnic_dev *dev)
val = BNX2_L2CTX_L2_STATUSB_NUM(sb_id); val = BNX2_L2CTX_L2_STATUSB_NUM(sb_id);
cnic_ctx_wr(dev, cid_addr, BNX2_L2CTX_HOST_BDIDX, val); cnic_ctx_wr(dev, cid_addr, BNX2_L2CTX_HOST_BDIDX, val);
rxbd = (struct rx_bd *) (udev->l2_ring + BCM_PAGE_SIZE); rxbd = udev->l2_ring + BCM_PAGE_SIZE;
for (i = 0; i < MAX_RX_DESC_CNT; i++, rxbd++) { for (i = 0; i < MAX_RX_DESC_CNT; i++, rxbd++) {
dma_addr_t buf_map; dma_addr_t buf_map;
int n = (i % cp->l2_rx_ring_size) + 1; int n = (i % cp->l2_rx_ring_size) + 1;

View file

@ -567,7 +567,7 @@ static void t3_process_tid_release_list(struct work_struct *work)
while (td->tid_release_list) { while (td->tid_release_list) {
struct t3c_tid_entry *p = td->tid_release_list; struct t3c_tid_entry *p = td->tid_release_list;
td->tid_release_list = (struct t3c_tid_entry *)p->ctx; td->tid_release_list = p->ctx;
spin_unlock_bh(&td->tid_release_lock); spin_unlock_bh(&td->tid_release_lock);
skb = alloc_skb(sizeof(struct cpl_tid_release), skb = alloc_skb(sizeof(struct cpl_tid_release),

View file

@ -167,7 +167,7 @@ cpdma_desc_pool_create(struct device *dev, u32 phys, u32 hw_addr,
} else { } else {
pool->cpumap = dma_alloc_coherent(dev, size, &pool->phys, pool->cpumap = dma_alloc_coherent(dev, size, &pool->phys,
GFP_KERNEL); GFP_KERNEL);
pool->iomap = (void __force __iomem *)pool->cpumap; pool->iomap = pool->cpumap;
pool->hw_addr = pool->phys; pool->hw_addr = pool->phys;
} }

View file

@ -326,15 +326,18 @@ static void load_csrs(struct lance_private *lp)
*/ */
static void cp_to_buf(const int type, void *to, const void *from, int len) static void cp_to_buf(const int type, void *to, const void *from, int len)
{ {
unsigned short *tp, *fp, clen; unsigned short *tp;
unsigned char *rtp, *rfp; const unsigned short *fp;
unsigned short clen;
unsigned char *rtp;
const unsigned char *rfp;
if (type == PMAD_LANCE) { if (type == PMAD_LANCE) {
memcpy(to, from, len); memcpy(to, from, len);
} else if (type == PMAX_LANCE) { } else if (type == PMAX_LANCE) {
clen = len >> 1; clen = len >> 1;
tp = (unsigned short *) to; tp = to;
fp = (unsigned short *) from; fp = from;
while (clen--) { while (clen--) {
*tp++ = *fp++; *tp++ = *fp++;
@ -342,8 +345,8 @@ static void cp_to_buf(const int type, void *to, const void *from, int len)
} }
clen = len & 1; clen = len & 1;
rtp = (unsigned char *) tp; rtp = tp;
rfp = (unsigned char *) fp; rfp = fp;
while (clen--) { while (clen--) {
*rtp++ = *rfp++; *rtp++ = *rfp++;
} }
@ -352,8 +355,8 @@ static void cp_to_buf(const int type, void *to, const void *from, int len)
* copy 16 Byte chunks * copy 16 Byte chunks
*/ */
clen = len >> 4; clen = len >> 4;
tp = (unsigned short *) to; tp = to;
fp = (unsigned short *) from; fp = from;
while (clen--) { while (clen--) {
*tp++ = *fp++; *tp++ = *fp++;
*tp++ = *fp++; *tp++ = *fp++;
@ -382,15 +385,18 @@ static void cp_to_buf(const int type, void *to, const void *from, int len)
static void cp_from_buf(const int type, void *to, const void *from, int len) static void cp_from_buf(const int type, void *to, const void *from, int len)
{ {
unsigned short *tp, *fp, clen; unsigned short *tp;
unsigned char *rtp, *rfp; const unsigned short *fp;
unsigned short clen;
unsigned char *rtp;
const unsigned char *rfp;
if (type == PMAD_LANCE) { if (type == PMAD_LANCE) {
memcpy(to, from, len); memcpy(to, from, len);
} else if (type == PMAX_LANCE) { } else if (type == PMAX_LANCE) {
clen = len >> 1; clen = len >> 1;
tp = (unsigned short *) to; tp = to;
fp = (unsigned short *) from; fp = from;
while (clen--) { while (clen--) {
*tp++ = *fp++; *tp++ = *fp++;
fp++; fp++;
@ -398,8 +404,8 @@ static void cp_from_buf(const int type, void *to, const void *from, int len)
clen = len & 1; clen = len & 1;
rtp = (unsigned char *) tp; rtp = tp;
rfp = (unsigned char *) fp; rfp = fp;
while (clen--) { while (clen--) {
*rtp++ = *rfp++; *rtp++ = *rfp++;
@ -410,8 +416,8 @@ static void cp_from_buf(const int type, void *to, const void *from, int len)
* copy 16 Byte chunks * copy 16 Byte chunks
*/ */
clen = len >> 4; clen = len >> 4;
tp = (unsigned short *) to; tp = to;
fp = (unsigned short *) from; fp = from;
while (clen--) { while (clen--) {
*tp++ = *fp++; *tp++ = *fp++;
*tp++ = *fp++; *tp++ = *fp++;

View file

@ -708,11 +708,11 @@ static int __init depca_hw_init (struct net_device *dev, struct device *device)
/* Tx & Rx descriptors (aligned to a quadword boundary) */ /* Tx & Rx descriptors (aligned to a quadword boundary) */
offset = (offset + DEPCA_ALIGN) & ~DEPCA_ALIGN; offset = (offset + DEPCA_ALIGN) & ~DEPCA_ALIGN;
lp->rx_ring = (struct depca_rx_desc __iomem *) (lp->sh_mem + offset); lp->rx_ring = lp->sh_mem + offset;
lp->rx_ring_offset = offset; lp->rx_ring_offset = offset;
offset += (sizeof(struct depca_rx_desc) * NUM_RX_DESC); offset += (sizeof(struct depca_rx_desc) * NUM_RX_DESC);
lp->tx_ring = (struct depca_tx_desc __iomem *) (lp->sh_mem + offset); lp->tx_ring = lp->sh_mem + offset;
lp->tx_ring_offset = offset; lp->tx_ring_offset = offset;
offset += (sizeof(struct depca_tx_desc) * NUM_TX_DESC); offset += (sizeof(struct depca_tx_desc) * NUM_TX_DESC);

View file

@ -221,13 +221,13 @@ rio_probe1 (struct pci_dev *pdev, const struct pci_device_id *ent)
ring_space = pci_alloc_consistent (pdev, TX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent (pdev, TX_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_iounmap; goto err_out_iounmap;
np->tx_ring = (struct netdev_desc *) ring_space; np->tx_ring = ring_space;
np->tx_ring_dma = ring_dma; np->tx_ring_dma = ring_dma;
ring_space = pci_alloc_consistent (pdev, RX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent (pdev, RX_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_unmap_tx; goto err_out_unmap_tx;
np->rx_ring = (struct netdev_desc *) ring_space; np->rx_ring = ring_space;
np->rx_ring_dma = ring_dma; np->rx_ring_dma = ring_dma;
/* Parse eeprom data */ /* Parse eeprom data */

View file

@ -331,7 +331,7 @@ struct ehea_eqe *ehea_poll_eq(struct ehea_eq *eq)
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&eq->spinlock, flags); spin_lock_irqsave(&eq->spinlock, flags);
eqe = (struct ehea_eqe *)hw_eqit_eq_get_inc_valid(&eq->hw_queue); eqe = hw_eqit_eq_get_inc_valid(&eq->hw_queue);
spin_unlock_irqrestore(&eq->spinlock, flags); spin_unlock_irqrestore(&eq->spinlock, flags);
return eqe; return eqe;

View file

@ -391,13 +391,13 @@ static int __devinit epic_init_one (struct pci_dev *pdev,
ring_space = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_iounmap; goto err_out_iounmap;
ep->tx_ring = (struct epic_tx_desc *)ring_space; ep->tx_ring = ring_space;
ep->tx_ring_dma = ring_dma; ep->tx_ring_dma = ring_dma;
ring_space = pci_alloc_consistent(pdev, RX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pdev, RX_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_unmap_tx; goto err_out_unmap_tx;
ep->rx_ring = (struct epic_rx_desc *)ring_space; ep->rx_ring = ring_space;
ep->rx_ring_dma = ring_dma; ep->rx_ring_dma = ring_dma;
if (dev->mem_start) { if (dev->mem_start) {

View file

@ -566,7 +566,7 @@ static int __devinit fealnx_init_one(struct pci_dev *pdev,
err = -ENOMEM; err = -ENOMEM;
goto err_out_free_dev; goto err_out_free_dev;
} }
np->rx_ring = (struct fealnx_desc *)ring_space; np->rx_ring = ring_space;
np->rx_ring_dma = ring_dma; np->rx_ring_dma = ring_dma;
ring_space = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma);
@ -574,7 +574,7 @@ static int __devinit fealnx_init_one(struct pci_dev *pdev,
err = -ENOMEM; err = -ENOMEM;
goto err_out_free_rx; goto err_out_free_rx;
} }
np->tx_ring = (struct fealnx_desc *)ring_space; np->tx_ring = ring_space;
np->tx_ring_dma = ring_dma; np->tx_ring_dma = ring_dma;
/* find the connected MII xcvrs */ /* find the connected MII xcvrs */

View file

@ -267,7 +267,7 @@ static int gfar_alloc_skb_resources(struct net_device *ndev)
for (i = 0; i < priv->num_tx_queues; i++) { for (i = 0; i < priv->num_tx_queues; i++) {
tx_queue = priv->tx_queue[i]; tx_queue = priv->tx_queue[i];
tx_queue->tx_bd_base = (struct txbd8 *) vaddr; tx_queue->tx_bd_base = vaddr;
tx_queue->tx_bd_dma_base = addr; tx_queue->tx_bd_dma_base = addr;
tx_queue->dev = ndev; tx_queue->dev = ndev;
/* enet DMA only understands physical addresses */ /* enet DMA only understands physical addresses */
@ -278,7 +278,7 @@ static int gfar_alloc_skb_resources(struct net_device *ndev)
/* Start the rx descriptor ring where the tx ring leaves off */ /* Start the rx descriptor ring where the tx ring leaves off */
for (i = 0; i < priv->num_rx_queues; i++) { for (i = 0; i < priv->num_rx_queues; i++) {
rx_queue = priv->rx_queue[i]; rx_queue = priv->rx_queue[i];
rx_queue->rx_bd_base = (struct rxbd8 *) vaddr; rx_queue->rx_bd_base = vaddr;
rx_queue->rx_bd_dma_base = addr; rx_queue->rx_bd_dma_base = addr;
rx_queue->dev = ndev; rx_queue->dev = ndev;
addr += sizeof (struct rxbd8) * rx_queue->rx_ring_size; addr += sizeof (struct rxbd8) * rx_queue->rx_ring_size;

View file

@ -648,13 +648,13 @@ static int __devinit hamachi_init_one (struct pci_dev *pdev,
ring_space = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_cleardev; goto err_out_cleardev;
hmp->tx_ring = (struct hamachi_desc *)ring_space; hmp->tx_ring = ring_space;
hmp->tx_ring_dma = ring_dma; hmp->tx_ring_dma = ring_dma;
ring_space = pci_alloc_consistent(pdev, RX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pdev, RX_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_unmap_tx; goto err_out_unmap_tx;
hmp->rx_ring = (struct hamachi_desc *)ring_space; hmp->rx_ring = ring_space;
hmp->rx_ring_dma = ring_dma; hmp->rx_ring_dma = ring_dma;
/* Check for options being passed in */ /* Check for options being passed in */

View file

@ -221,7 +221,7 @@ static int __devinit mace_probe(struct platform_device *pdev)
SET_NETDEV_DEV(dev, &pdev->dev); SET_NETDEV_DEV(dev, &pdev->dev);
dev->base_addr = (u32)MACE_BASE; dev->base_addr = (u32)MACE_BASE;
mp->mace = (volatile struct mace *) MACE_BASE; mp->mace = MACE_BASE;
dev->irq = IRQ_MAC_MACE; dev->irq = IRQ_MAC_MACE;
mp->dma_intr = IRQ_MAC_MACE_DMA; mp->dma_intr = IRQ_MAC_MACE_DMA;

View file

@ -859,7 +859,7 @@ int mlx4_en_config_rss_steer(struct mlx4_en_priv *priv)
priv->rx_ring[0].cqn, &context); priv->rx_ring[0].cqn, &context);
ptr = ((void *) &context) + 0x3c; ptr = ((void *) &context) + 0x3c;
rss_context = (struct mlx4_en_rss_context *) ptr; rss_context = ptr;
rss_context->base_qpn = cpu_to_be32(ilog2(priv->rx_ring_num) << 24 | rss_context->base_qpn = cpu_to_be32(ilog2(priv->rx_ring_num) << 24 |
(rss_map->base_qpn)); (rss_map->base_qpn));
rss_context->default_qpn = cpu_to_be32(rss_map->base_qpn); rss_context->default_qpn = cpu_to_be32(rss_map->base_qpn);

View file

@ -238,8 +238,7 @@ static u32 mlx4_en_free_tx_desc(struct mlx4_en_priv *priv,
} else { } else {
if (!tx_info->inl) { if (!tx_info->inl) {
if ((void *) data >= end) { if ((void *) data >= end) {
data = (struct mlx4_wqe_data_seg *) data = ring->buf + ((void *)data - end);
(ring->buf + ((void *) data - end));
} }
if (tx_info->linear) { if (tx_info->linear) {
@ -253,7 +252,7 @@ static u32 mlx4_en_free_tx_desc(struct mlx4_en_priv *priv,
for (i = 0; i < frags; i++) { for (i = 0; i < frags; i++) {
/* Check for wraparound before unmapping */ /* Check for wraparound before unmapping */
if ((void *) data >= end) if ((void *) data >= end)
data = (struct mlx4_wqe_data_seg *) ring->buf; data = ring->buf;
frag = &skb_shinfo(skb)->frags[i]; frag = &skb_shinfo(skb)->frags[i];
pci_unmap_page(mdev->pdev, pci_unmap_page(mdev->pdev,
(dma_addr_t) be64_to_cpu(data->addr), (dma_addr_t) be64_to_cpu(data->addr),

View file

@ -163,7 +163,7 @@ nx_fw_cmd_create_rx_ctx(struct netxen_adapter *adapter)
rq_size, &hostrq_phys_addr); rq_size, &hostrq_phys_addr);
if (addr == NULL) if (addr == NULL)
return -ENOMEM; return -ENOMEM;
prq = (nx_hostrq_rx_ctx_t *)addr; prq = addr;
addr = pci_alloc_consistent(adapter->pdev, addr = pci_alloc_consistent(adapter->pdev,
rsp_size, &cardrsp_phys_addr); rsp_size, &cardrsp_phys_addr);
@ -171,7 +171,7 @@ nx_fw_cmd_create_rx_ctx(struct netxen_adapter *adapter)
err = -ENOMEM; err = -ENOMEM;
goto out_free_rq; goto out_free_rq;
} }
prsp = (nx_cardrsp_rx_ctx_t *)addr; prsp = addr;
prq->host_rsp_dma_addr = cpu_to_le64(cardrsp_phys_addr); prq->host_rsp_dma_addr = cpu_to_le64(cardrsp_phys_addr);
@ -318,10 +318,10 @@ nx_fw_cmd_create_tx_ctx(struct netxen_adapter *adapter)
} }
memset(rq_addr, 0, rq_size); memset(rq_addr, 0, rq_size);
prq = (nx_hostrq_tx_ctx_t *)rq_addr; prq = rq_addr;
memset(rsp_addr, 0, rsp_size); memset(rsp_addr, 0, rsp_size);
prsp = (nx_cardrsp_tx_ctx_t *)rsp_addr; prsp = rsp_addr;
prq->host_rsp_dma_addr = cpu_to_le64(rsp_phys_addr); prq->host_rsp_dma_addr = cpu_to_le64(rsp_phys_addr);
@ -629,7 +629,7 @@ int netxen_alloc_hw_resources(struct netxen_adapter *adapter)
} }
memset(addr, 0, sizeof(struct netxen_ring_ctx)); memset(addr, 0, sizeof(struct netxen_ring_ctx));
recv_ctx->hwctx = (struct netxen_ring_ctx *)addr; recv_ctx->hwctx = addr;
recv_ctx->hwctx->ctx_id = cpu_to_le32(port); recv_ctx->hwctx->ctx_id = cpu_to_le32(port);
recv_ctx->hwctx->cmd_consumer_offset = recv_ctx->hwctx->cmd_consumer_offset =
cpu_to_le64(recv_ctx->phys_addr + cpu_to_le64(recv_ctx->phys_addr +
@ -648,7 +648,7 @@ int netxen_alloc_hw_resources(struct netxen_adapter *adapter)
goto err_out_free; goto err_out_free;
} }
tx_ring->desc_head = (struct cmd_desc_type0 *)addr; tx_ring->desc_head = addr;
for (ring = 0; ring < adapter->max_rds_rings; ring++) { for (ring = 0; ring < adapter->max_rds_rings; ring++) {
rds_ring = &recv_ctx->rds_rings[ring]; rds_ring = &recv_ctx->rds_rings[ring];
@ -662,7 +662,7 @@ int netxen_alloc_hw_resources(struct netxen_adapter *adapter)
err = -ENOMEM; err = -ENOMEM;
goto err_out_free; goto err_out_free;
} }
rds_ring->desc_head = (struct rcv_desc *)addr; rds_ring->desc_head = addr;
if (NX_IS_REVISION_P2(adapter->ahw.revision_id)) if (NX_IS_REVISION_P2(adapter->ahw.revision_id))
rds_ring->crb_rcv_producer = rds_ring->crb_rcv_producer =
@ -683,7 +683,7 @@ int netxen_alloc_hw_resources(struct netxen_adapter *adapter)
err = -ENOMEM; err = -ENOMEM;
goto err_out_free; goto err_out_free;
} }
sds_ring->desc_head = (struct status_desc *)addr; sds_ring->desc_head = addr;
if (NX_IS_REVISION_P2(adapter->ahw.revision_id)) { if (NX_IS_REVISION_P2(adapter->ahw.revision_id)) {
sds_ring->crb_sts_consumer = sds_ring->crb_sts_consumer =

View file

@ -502,7 +502,7 @@ static int add_del_hash_entry(struct pxa168_eth_private *pep,
* Pick the appropriate table, start scanning for free/reusable * Pick the appropriate table, start scanning for free/reusable
* entries at the index obtained by hashing the specified MAC address * entries at the index obtained by hashing the specified MAC address
*/ */
start = (struct addr_table_entry *)(pep->htpr); start = pep->htpr;
entry = start + hash_function(mac_addr); entry = start + hash_function(mac_addr);
for (i = 0; i < HOP_NUMBER; i++) { for (i = 0; i < HOP_NUMBER; i++) {
if (!(le32_to_cpu(entry->lo) & HASH_ENTRY_VALID)) { if (!(le32_to_cpu(entry->lo) & HASH_ENTRY_VALID)) {

View file

@ -2873,7 +2873,7 @@ static int ql_alloc_mem_resources(struct ql3_adapter *qdev)
PAGE_SIZE, &qdev->shadow_reg_phy_addr); PAGE_SIZE, &qdev->shadow_reg_phy_addr);
if (qdev->shadow_reg_virt_addr != NULL) { if (qdev->shadow_reg_virt_addr != NULL) {
qdev->preq_consumer_index = (u16 *) qdev->shadow_reg_virt_addr; qdev->preq_consumer_index = qdev->shadow_reg_virt_addr;
qdev->req_consumer_index_phy_addr_high = qdev->req_consumer_index_phy_addr_high =
MS_64BITS(qdev->shadow_reg_phy_addr); MS_64BITS(qdev->shadow_reg_phy_addr);
qdev->req_consumer_index_phy_addr_low = qdev->req_consumer_index_phy_addr_low =
@ -3114,8 +3114,7 @@ static int ql_adapter_initialize(struct ql3_adapter *qdev)
qdev->small_buf_release_cnt = 8; qdev->small_buf_release_cnt = 8;
qdev->lrg_buf_q_producer_index = qdev->num_lbufq_entries - 1; qdev->lrg_buf_q_producer_index = qdev->num_lbufq_entries - 1;
qdev->lrg_buf_release_cnt = 8; qdev->lrg_buf_release_cnt = 8;
qdev->lrg_buf_next_free = qdev->lrg_buf_next_free = qdev->lrg_buf_q_virt_addr;
(struct bufq_addr_element *)qdev->lrg_buf_q_virt_addr;
qdev->small_buf_index = 0; qdev->small_buf_index = 0;
qdev->lrg_buf_index = 0; qdev->lrg_buf_index = 0;
qdev->lrg_buf_free_count = 0; qdev->lrg_buf_free_count = 0;

View file

@ -126,7 +126,7 @@ int qlcnic_fw_cmd_get_minidump_temp(struct qlcnic_adapter *adapter)
err = -EIO; err = -EIO;
goto error; goto error;
} }
tmp_tmpl = (struct qlcnic_dump_template_hdr *) tmp_addr; tmp_tmpl = tmp_addr;
csum = qlcnic_temp_checksum((uint32_t *) tmp_addr, temp_size); csum = qlcnic_temp_checksum((uint32_t *) tmp_addr, temp_size);
if (csum) { if (csum) {
dev_err(&adapter->pdev->dev, dev_err(&adapter->pdev->dev,
@ -139,7 +139,7 @@ int qlcnic_fw_cmd_get_minidump_temp(struct qlcnic_adapter *adapter)
err = -EIO; err = -EIO;
goto error; goto error;
} }
tmp_buf = (u32 *) tmp_addr; tmp_buf = tmp_addr;
template = (u32 *) ahw->fw_dump.tmpl_hdr; template = (u32 *) ahw->fw_dump.tmpl_hdr;
for (i = 0; i < temp_size/sizeof(u32); i++) for (i = 0; i < temp_size/sizeof(u32); i++)
*template++ = __le32_to_cpu(*tmp_buf++); *template++ = __le32_to_cpu(*tmp_buf++);
@ -214,7 +214,7 @@ qlcnic_fw_cmd_create_rx_ctx(struct qlcnic_adapter *adapter)
&hostrq_phys_addr, GFP_KERNEL); &hostrq_phys_addr, GFP_KERNEL);
if (addr == NULL) if (addr == NULL)
return -ENOMEM; return -ENOMEM;
prq = (struct qlcnic_hostrq_rx_ctx *)addr; prq = addr;
addr = dma_alloc_coherent(&adapter->pdev->dev, rsp_size, addr = dma_alloc_coherent(&adapter->pdev->dev, rsp_size,
&cardrsp_phys_addr, GFP_KERNEL); &cardrsp_phys_addr, GFP_KERNEL);
@ -222,7 +222,7 @@ qlcnic_fw_cmd_create_rx_ctx(struct qlcnic_adapter *adapter)
err = -ENOMEM; err = -ENOMEM;
goto out_free_rq; goto out_free_rq;
} }
prsp = (struct qlcnic_cardrsp_rx_ctx *)addr; prsp = addr;
prq->host_rsp_dma_addr = cpu_to_le64(cardrsp_phys_addr); prq->host_rsp_dma_addr = cpu_to_le64(cardrsp_phys_addr);
@ -380,10 +380,10 @@ qlcnic_fw_cmd_create_tx_ctx(struct qlcnic_adapter *adapter)
} }
memset(rq_addr, 0, rq_size); memset(rq_addr, 0, rq_size);
prq = (struct qlcnic_hostrq_tx_ctx *)rq_addr; prq = rq_addr;
memset(rsp_addr, 0, rsp_size); memset(rsp_addr, 0, rsp_size);
prsp = (struct qlcnic_cardrsp_tx_ctx *)rsp_addr; prsp = rsp_addr;
prq->host_rsp_dma_addr = cpu_to_le64(rsp_phys_addr); prq->host_rsp_dma_addr = cpu_to_le64(rsp_phys_addr);
@ -493,7 +493,7 @@ int qlcnic_alloc_hw_resources(struct qlcnic_adapter *adapter)
goto err_out_free; goto err_out_free;
} }
tx_ring->desc_head = (struct cmd_desc_type0 *)addr; tx_ring->desc_head = addr;
for (ring = 0; ring < adapter->max_rds_rings; ring++) { for (ring = 0; ring < adapter->max_rds_rings; ring++) {
rds_ring = &recv_ctx->rds_rings[ring]; rds_ring = &recv_ctx->rds_rings[ring];
@ -506,7 +506,7 @@ int qlcnic_alloc_hw_resources(struct qlcnic_adapter *adapter)
err = -ENOMEM; err = -ENOMEM;
goto err_out_free; goto err_out_free;
} }
rds_ring->desc_head = (struct rcv_desc *)addr; rds_ring->desc_head = addr;
} }
@ -522,7 +522,7 @@ int qlcnic_alloc_hw_resources(struct qlcnic_adapter *adapter)
err = -ENOMEM; err = -ENOMEM;
goto err_out_free; goto err_out_free;
} }
sds_ring->desc_head = (struct status_desc *)addr; sds_ring->desc_head = addr;
} }
return 0; return 0;
@ -662,7 +662,7 @@ int qlcnic_get_nic_info(struct qlcnic_adapter *adapter,
return -ENOMEM; return -ENOMEM;
memset(nic_info_addr, 0, nic_size); memset(nic_info_addr, 0, nic_size);
nic_info = (struct qlcnic_info *) nic_info_addr; nic_info = nic_info_addr;
err = qlcnic_issue_cmd(adapter, err = qlcnic_issue_cmd(adapter,
adapter->ahw->pci_func, adapter->ahw->pci_func,
adapter->fw_hal_version, adapter->fw_hal_version,
@ -720,7 +720,7 @@ int qlcnic_set_nic_info(struct qlcnic_adapter *adapter, struct qlcnic_info *nic)
return -ENOMEM; return -ENOMEM;
memset(nic_info_addr, 0, nic_size); memset(nic_info_addr, 0, nic_size);
nic_info = (struct qlcnic_info *)nic_info_addr; nic_info = nic_info_addr;
nic_info->pci_func = cpu_to_le16(nic->pci_func); nic_info->pci_func = cpu_to_le16(nic->pci_func);
nic_info->op_mode = cpu_to_le16(nic->op_mode); nic_info->op_mode = cpu_to_le16(nic->op_mode);
@ -769,7 +769,7 @@ int qlcnic_get_pci_info(struct qlcnic_adapter *adapter,
return -ENOMEM; return -ENOMEM;
memset(pci_info_addr, 0, pci_size); memset(pci_info_addr, 0, pci_size);
npar = (struct qlcnic_pci_info *) pci_info_addr; npar = pci_info_addr;
err = qlcnic_issue_cmd(adapter, err = qlcnic_issue_cmd(adapter,
adapter->ahw->pci_func, adapter->ahw->pci_func,
adapter->fw_hal_version, adapter->fw_hal_version,
@ -877,7 +877,7 @@ int qlcnic_get_port_stats(struct qlcnic_adapter *adapter, const u8 func,
QLCNIC_CDRP_CMD_GET_ESWITCH_STATS); QLCNIC_CDRP_CMD_GET_ESWITCH_STATS);
if (!err) { if (!err) {
stats = (struct __qlcnic_esw_statistics *)stats_addr; stats = stats_addr;
esw_stats->context_id = le16_to_cpu(stats->context_id); esw_stats->context_id = le16_to_cpu(stats->context_id);
esw_stats->version = le16_to_cpu(stats->version); esw_stats->version = le16_to_cpu(stats->version);
esw_stats->size = le16_to_cpu(stats->size); esw_stats->size = le16_to_cpu(stats->size);

View file

@ -996,7 +996,7 @@ qlcnic_get_dump_data(struct net_device *netdev, struct ethtool_dump *dump,
/* Copy template header first */ /* Copy template header first */
copy_sz = fw_dump->tmpl_hdr->size; copy_sz = fw_dump->tmpl_hdr->size;
hdr_ptr = (u32 *) fw_dump->tmpl_hdr; hdr_ptr = (u32 *) fw_dump->tmpl_hdr;
data = (u32 *) buffer; data = buffer;
for (i = 0; i < copy_sz/sizeof(u32); i++) for (i = 0; i < copy_sz/sizeof(u32); i++)
*data++ = cpu_to_le32(*hdr_ptr++); *data++ = cpu_to_le32(*hdr_ptr++);

View file

@ -1673,8 +1673,7 @@ int qlcnic_dump_fw(struct qlcnic_adapter *adapter)
tmpl_hdr->sys_info[1] = adapter->fw_version; tmpl_hdr->sys_info[1] = adapter->fw_version;
for (i = 0; i < no_entries; i++) { for (i = 0; i < no_entries; i++) {
entry = (struct qlcnic_dump_entry *) ((void *) tmpl_hdr + entry = (void *)tmpl_hdr + entry_offset;
entry_offset);
if (!(entry->hdr.mask & tmpl_hdr->drv_cap_mask)) { if (!(entry->hdr.mask & tmpl_hdr->drv_cap_mask)) {
entry->hdr.flags |= QLCNIC_DUMP_SKIP; entry->hdr.flags |= QLCNIC_DUMP_SKIP;
entry_offset += entry->hdr.offset; entry_offset += entry->hdr.offset;

View file

@ -1861,7 +1861,7 @@ qlcnic_alloc_lb_filters_mem(struct qlcnic_adapter *adapter)
return; return;
adapter->fhash.fmax = QLCNIC_LB_MAX_FILTERS; adapter->fhash.fmax = QLCNIC_LB_MAX_FILTERS;
adapter->fhash.fhead = (struct hlist_head *)head; adapter->fhash.fhead = head;
for (i = 0; i < adapter->fhash.fmax; i++) for (i = 0; i < adapter->fhash.fmax; i++)
INIT_HLIST_HEAD(&adapter->fhash.fhead[i]); INIT_HLIST_HEAD(&adapter->fhash.fhead[i]);

View file

@ -3096,7 +3096,7 @@ static int ql_start_rx_ring(struct ql_adapter *qdev, struct rx_ring *rx_ring)
if (rx_ring->lbq_len) { if (rx_ring->lbq_len) {
cqicb->flags |= FLAGS_LL; /* Load lbq values */ cqicb->flags |= FLAGS_LL; /* Load lbq values */
tmp = (u64)rx_ring->lbq_base_dma; tmp = (u64)rx_ring->lbq_base_dma;
base_indirect_ptr = (__le64 *) rx_ring->lbq_base_indirect; base_indirect_ptr = rx_ring->lbq_base_indirect;
page_entries = 0; page_entries = 0;
do { do {
*base_indirect_ptr = cpu_to_le64(tmp); *base_indirect_ptr = cpu_to_le64(tmp);
@ -3120,7 +3120,7 @@ static int ql_start_rx_ring(struct ql_adapter *qdev, struct rx_ring *rx_ring)
if (rx_ring->sbq_len) { if (rx_ring->sbq_len) {
cqicb->flags |= FLAGS_LS; /* Load sbq values */ cqicb->flags |= FLAGS_LS; /* Load sbq values */
tmp = (u64)rx_ring->sbq_base_dma; tmp = (u64)rx_ring->sbq_base_dma;
base_indirect_ptr = (__le64 *) rx_ring->sbq_base_indirect; base_indirect_ptr = rx_ring->sbq_base_indirect;
page_entries = 0; page_entries = 0;
do { do {
*base_indirect_ptr = cpu_to_le64(tmp); *base_indirect_ptr = cpu_to_le64(tmp);

View file

@ -841,7 +841,7 @@ static int init_shared_mem(struct s2io_nic *nic)
tmp_p_addr = ring->rx_blocks[j].block_dma_addr; tmp_p_addr = ring->rx_blocks[j].block_dma_addr;
tmp_p_addr_next = ring->rx_blocks[next].block_dma_addr; tmp_p_addr_next = ring->rx_blocks[next].block_dma_addr;
pre_rxd_blk = (struct RxD_block *)tmp_v_addr; pre_rxd_blk = tmp_v_addr;
pre_rxd_blk->reserved_2_pNext_RxD_block = pre_rxd_blk->reserved_2_pNext_RxD_block =
(unsigned long)tmp_v_addr_next; (unsigned long)tmp_v_addr_next;
pre_rxd_blk->pNext_RxD_Blk_physical = pre_rxd_blk->pNext_RxD_Blk_physical =
@ -918,7 +918,7 @@ static int init_shared_mem(struct s2io_nic *nic)
mac_control->stats_mem_sz = size; mac_control->stats_mem_sz = size;
tmp_v_addr = mac_control->stats_mem; tmp_v_addr = mac_control->stats_mem;
mac_control->stats_info = (struct stat_block *)tmp_v_addr; mac_control->stats_info = tmp_v_addr;
memset(tmp_v_addr, 0, size); memset(tmp_v_addr, 0, size);
DBG_PRINT(INIT_DBG, "%s: Ring Mem PHY: 0x%llx\n", DBG_PRINT(INIT_DBG, "%s: Ring Mem PHY: 0x%llx\n",
dev_name(&nic->pdev->dev), (unsigned long long)tmp_p_addr); dev_name(&nic->pdev->dev), (unsigned long long)tmp_p_addr);
@ -2439,7 +2439,7 @@ static void free_tx_buffers(struct s2io_nic *nic)
spin_lock_irqsave(&fifo->tx_lock, flags); spin_lock_irqsave(&fifo->tx_lock, flags);
for (j = 0; j < tx_cfg->fifo_len; j++) { for (j = 0; j < tx_cfg->fifo_len; j++) {
txdp = (struct TxD *)fifo->list_info[j].list_virt_addr; txdp = fifo->list_info[j].list_virt_addr;
skb = s2io_txdl_getskb(&mac_control->fifos[i], txdp, j); skb = s2io_txdl_getskb(&mac_control->fifos[i], txdp, j);
if (skb) { if (skb) {
swstats->mem_freed += skb->truesize; swstats->mem_freed += skb->truesize;
@ -3075,8 +3075,7 @@ static void tx_intr_handler(struct fifo_info *fifo_data)
get_info = fifo_data->tx_curr_get_info; get_info = fifo_data->tx_curr_get_info;
memcpy(&put_info, &fifo_data->tx_curr_put_info, sizeof(put_info)); memcpy(&put_info, &fifo_data->tx_curr_put_info, sizeof(put_info));
txdlp = (struct TxD *) txdlp = fifo_data->list_info[get_info.offset].list_virt_addr;
fifo_data->list_info[get_info.offset].list_virt_addr;
while ((!(txdlp->Control_1 & TXD_LIST_OWN_XENA)) && while ((!(txdlp->Control_1 & TXD_LIST_OWN_XENA)) &&
(get_info.offset != put_info.offset) && (get_info.offset != put_info.offset) &&
(txdlp->Host_Control)) { (txdlp->Host_Control)) {
@ -3129,8 +3128,7 @@ static void tx_intr_handler(struct fifo_info *fifo_data)
get_info.offset++; get_info.offset++;
if (get_info.offset == get_info.fifo_len + 1) if (get_info.offset == get_info.fifo_len + 1)
get_info.offset = 0; get_info.offset = 0;
txdlp = (struct TxD *) txdlp = fifo_data->list_info[get_info.offset].list_virt_addr;
fifo_data->list_info[get_info.offset].list_virt_addr;
fifo_data->tx_curr_get_info.offset = get_info.offset; fifo_data->tx_curr_get_info.offset = get_info.offset;
} }
@ -4163,7 +4161,7 @@ static netdev_tx_t s2io_xmit(struct sk_buff *skb, struct net_device *dev)
put_off = (u16)fifo->tx_curr_put_info.offset; put_off = (u16)fifo->tx_curr_put_info.offset;
get_off = (u16)fifo->tx_curr_get_info.offset; get_off = (u16)fifo->tx_curr_get_info.offset;
txdp = (struct TxD *)fifo->list_info[put_off].list_virt_addr; txdp = fifo->list_info[put_off].list_virt_addr;
queue_len = fifo->tx_curr_put_info.fifo_len + 1; queue_len = fifo->tx_curr_put_info.fifo_len + 1;
/* Avoid "put" pointer going beyond "get" pointer */ /* Avoid "put" pointer going beyond "get" pointer */
@ -7972,9 +7970,7 @@ s2io_init_nic(struct pci_dev *pdev, const struct pci_device_id *pre)
/* Initializing the BAR1 address as the start of the FIFO pointer. */ /* Initializing the BAR1 address as the start of the FIFO pointer. */
for (j = 0; j < MAX_TX_FIFOS; j++) { for (j = 0; j < MAX_TX_FIFOS; j++) {
mac_control->tx_FIFO_start[j] = mac_control->tx_FIFO_start[j] = sp->bar1 + (j * 0x00020000);
(struct TxFIFO_element __iomem *)
(sp->bar1 + (j * 0x00020000));
} }
/* Driver entry points */ /* Driver entry points */

View file

@ -400,7 +400,7 @@ static int siena_try_update_nic_stats(struct efx_nic *efx)
u64 generation_end; u64 generation_end;
mac_stats = &efx->mac_stats; mac_stats = &efx->mac_stats;
dma_stats = (u64 *)efx->stats_buffer.addr; dma_stats = efx->stats_buffer.addr;
generation_end = dma_stats[MC_CMD_MAC_GENERATION_END]; generation_end = dma_stats[MC_CMD_MAC_GENERATION_END];
if (generation_end == STATS_GENERATION_INVALID) if (generation_end == STATS_GENERATION_INVALID)

View file

@ -482,7 +482,7 @@ static int __devinit sis900_probe(struct pci_dev *pci_dev,
ret = -ENOMEM; ret = -ENOMEM;
goto err_out_cleardev; goto err_out_cleardev;
} }
sis_priv->tx_ring = (BufferDesc *)ring_space; sis_priv->tx_ring = ring_space;
sis_priv->tx_ring_dma = ring_dma; sis_priv->tx_ring_dma = ring_dma;
ring_space = pci_alloc_consistent(pci_dev, RX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pci_dev, RX_TOTAL_SIZE, &ring_dma);
@ -490,7 +490,7 @@ static int __devinit sis900_probe(struct pci_dev *pci_dev,
ret = -ENOMEM; ret = -ENOMEM;
goto err_unmap_tx; goto err_unmap_tx;
} }
sis_priv->rx_ring = (BufferDesc *)ring_space; sis_priv->rx_ring = ring_space;
sis_priv->rx_ring_dma = ring_dma; sis_priv->rx_ring_dma = ring_dma;
/* The SiS900-specific entries in the device structure. */ /* The SiS900-specific entries in the device structure. */

View file

@ -418,7 +418,7 @@ static irqreturn_t madgemc_interrupt(int irq, void *dev_id)
return IRQ_NONE; return IRQ_NONE;
} }
dev = (struct net_device *)dev_id; dev = dev_id;
/* Make sure its really us. -- the Madge way */ /* Make sure its really us. -- the Madge way */
pending = inb(dev->base_addr + MC_CONTROL_REG0); pending = inb(dev->base_addr + MC_CONTROL_REG0);

View file

@ -2367,7 +2367,7 @@ typhoon_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
dev->irq = pdev->irq; dev->irq = pdev->irq;
tp = netdev_priv(dev); tp = netdev_priv(dev);
tp->shared = (struct typhoon_shared *) shared; tp->shared = shared;
tp->shared_dma = shared_dma; tp->shared_dma = shared_dma;
tp->pdev = pdev; tp->pdev = pdev;
tp->tx_pdev = pdev; tp->tx_pdev = pdev;

View file

@ -582,7 +582,7 @@ __vxge_hw_device_toc_get(void __iomem *bar0)
goto exit; goto exit;
val64 = readq(&legacy_reg->toc_first_pointer); val64 = readq(&legacy_reg->toc_first_pointer);
toc = (struct vxge_hw_toc_reg __iomem *)(bar0+val64); toc = bar0 + val64;
exit: exit:
return toc; return toc;
} }
@ -600,7 +600,7 @@ __vxge_hw_device_reg_addr_get(struct __vxge_hw_device *hldev)
u32 i; u32 i;
enum vxge_hw_status status = VXGE_HW_OK; enum vxge_hw_status status = VXGE_HW_OK;
hldev->legacy_reg = (struct vxge_hw_legacy_reg __iomem *)hldev->bar0; hldev->legacy_reg = hldev->bar0;
hldev->toc_reg = __vxge_hw_device_toc_get(hldev->bar0); hldev->toc_reg = __vxge_hw_device_toc_get(hldev->bar0);
if (hldev->toc_reg == NULL) { if (hldev->toc_reg == NULL) {
@ -609,39 +609,31 @@ __vxge_hw_device_reg_addr_get(struct __vxge_hw_device *hldev)
} }
val64 = readq(&hldev->toc_reg->toc_common_pointer); val64 = readq(&hldev->toc_reg->toc_common_pointer);
hldev->common_reg = hldev->common_reg = hldev->bar0 + val64;
(struct vxge_hw_common_reg __iomem *)(hldev->bar0 + val64);
val64 = readq(&hldev->toc_reg->toc_mrpcim_pointer); val64 = readq(&hldev->toc_reg->toc_mrpcim_pointer);
hldev->mrpcim_reg = hldev->mrpcim_reg = hldev->bar0 + val64;
(struct vxge_hw_mrpcim_reg __iomem *)(hldev->bar0 + val64);
for (i = 0; i < VXGE_HW_TITAN_SRPCIM_REG_SPACES; i++) { for (i = 0; i < VXGE_HW_TITAN_SRPCIM_REG_SPACES; i++) {
val64 = readq(&hldev->toc_reg->toc_srpcim_pointer[i]); val64 = readq(&hldev->toc_reg->toc_srpcim_pointer[i]);
hldev->srpcim_reg[i] = hldev->srpcim_reg[i] = hldev->bar0 + val64;
(struct vxge_hw_srpcim_reg __iomem *)
(hldev->bar0 + val64);
} }
for (i = 0; i < VXGE_HW_TITAN_VPMGMT_REG_SPACES; i++) { for (i = 0; i < VXGE_HW_TITAN_VPMGMT_REG_SPACES; i++) {
val64 = readq(&hldev->toc_reg->toc_vpmgmt_pointer[i]); val64 = readq(&hldev->toc_reg->toc_vpmgmt_pointer[i]);
hldev->vpmgmt_reg[i] = hldev->vpmgmt_reg[i] = hldev->bar0 + val64;
(struct vxge_hw_vpmgmt_reg __iomem *)(hldev->bar0 + val64);
} }
for (i = 0; i < VXGE_HW_TITAN_VPATH_REG_SPACES; i++) { for (i = 0; i < VXGE_HW_TITAN_VPATH_REG_SPACES; i++) {
val64 = readq(&hldev->toc_reg->toc_vpath_pointer[i]); val64 = readq(&hldev->toc_reg->toc_vpath_pointer[i]);
hldev->vpath_reg[i] = hldev->vpath_reg[i] = hldev->bar0 + val64;
(struct vxge_hw_vpath_reg __iomem *)
(hldev->bar0 + val64);
} }
val64 = readq(&hldev->toc_reg->toc_kdfc); val64 = readq(&hldev->toc_reg->toc_kdfc);
switch (VXGE_HW_TOC_GET_KDFC_INITIAL_BIR(val64)) { switch (VXGE_HW_TOC_GET_KDFC_INITIAL_BIR(val64)) {
case 0: case 0:
hldev->kdfc = (u8 __iomem *)(hldev->bar0 + hldev->kdfc = hldev->bar0 + VXGE_HW_TOC_GET_KDFC_INITIAL_OFFSET(val64) ;
VXGE_HW_TOC_GET_KDFC_INITIAL_OFFSET(val64));
break; break;
default: default:
break; break;
@ -1024,7 +1016,7 @@ vxge_hw_device_hw_info_get(void __iomem *bar0,
} }
val64 = readq(&toc->toc_common_pointer); val64 = readq(&toc->toc_common_pointer);
common_reg = (struct vxge_hw_common_reg __iomem *)(bar0 + val64); common_reg = bar0 + val64;
status = __vxge_hw_device_vpath_reset_in_prog_check( status = __vxge_hw_device_vpath_reset_in_prog_check(
(u64 __iomem *)&common_reg->vpath_rst_in_prog); (u64 __iomem *)&common_reg->vpath_rst_in_prog);
@ -1044,8 +1036,7 @@ vxge_hw_device_hw_info_get(void __iomem *bar0,
val64 = readq(&toc->toc_vpmgmt_pointer[i]); val64 = readq(&toc->toc_vpmgmt_pointer[i]);
vpmgmt_reg = (struct vxge_hw_vpmgmt_reg __iomem *) vpmgmt_reg = bar0 + val64;
(bar0 + val64);
hw_info->func_id = __vxge_hw_vpath_func_id_get(vpmgmt_reg); hw_info->func_id = __vxge_hw_vpath_func_id_get(vpmgmt_reg);
if (__vxge_hw_device_access_rights_get(hw_info->host_type, if (__vxge_hw_device_access_rights_get(hw_info->host_type,
@ -1054,8 +1045,7 @@ vxge_hw_device_hw_info_get(void __iomem *bar0,
val64 = readq(&toc->toc_mrpcim_pointer); val64 = readq(&toc->toc_mrpcim_pointer);
mrpcim_reg = (struct vxge_hw_mrpcim_reg __iomem *) mrpcim_reg = bar0 + val64;
(bar0 + val64);
writeq(0, &mrpcim_reg->xgmac_gen_fw_memo_mask); writeq(0, &mrpcim_reg->xgmac_gen_fw_memo_mask);
wmb(); wmb();
@ -1064,8 +1054,7 @@ vxge_hw_device_hw_info_get(void __iomem *bar0,
val64 = readq(&toc->toc_vpath_pointer[i]); val64 = readq(&toc->toc_vpath_pointer[i]);
spin_lock_init(&vpath.lock); spin_lock_init(&vpath.lock);
vpath.vp_reg = (struct vxge_hw_vpath_reg __iomem *) vpath.vp_reg = bar0 + val64;
(bar0 + val64);
vpath.vp_open = VXGE_HW_VP_NOT_OPEN; vpath.vp_open = VXGE_HW_VP_NOT_OPEN;
status = __vxge_hw_vpath_pci_func_mode_get(&vpath, hw_info); status = __vxge_hw_vpath_pci_func_mode_get(&vpath, hw_info);
@ -1088,8 +1077,7 @@ vxge_hw_device_hw_info_get(void __iomem *bar0,
continue; continue;
val64 = readq(&toc->toc_vpath_pointer[i]); val64 = readq(&toc->toc_vpath_pointer[i]);
vpath.vp_reg = (struct vxge_hw_vpath_reg __iomem *) vpath.vp_reg = bar0 + val64;
(bar0 + val64);
vpath.vp_open = VXGE_HW_VP_NOT_OPEN; vpath.vp_open = VXGE_HW_VP_NOT_OPEN;
status = __vxge_hw_vpath_addr_get(&vpath, status = __vxge_hw_vpath_addr_get(&vpath,
@ -2140,8 +2128,7 @@ __vxge_hw_ring_mempool_item_alloc(struct vxge_hw_mempool *mempoolh,
memblock_index, item, memblock_index, item,
&memblock_item_idx); &memblock_item_idx);
rxdp = (struct vxge_hw_ring_rxd_1 *) rxdp = ring->channel.reserve_arr[reserve_index];
ring->channel.reserve_arr[reserve_index];
uld_priv = ((u8 *)rxdblock_priv + ring->rxd_priv_size * i); uld_priv = ((u8 *)rxdblock_priv + ring->rxd_priv_size * i);
@ -4880,8 +4867,7 @@ vxge_hw_vpath_open(struct __vxge_hw_device *hldev,
goto vpath_open_exit8; goto vpath_open_exit8;
} }
vpath->hw_stats = (struct vxge_hw_vpath_stats_hw_info *)vpath-> vpath->hw_stats = vpath->stats_block->memblock;
stats_block->memblock;
memset(vpath->hw_stats, 0, memset(vpath->hw_stats, 0,
sizeof(struct vxge_hw_vpath_stats_hw_info)); sizeof(struct vxge_hw_vpath_stats_hw_info));

View file

@ -1309,7 +1309,7 @@ enum vxge_hw_status vxge_hw_ring_rxd_next_completed(
vxge_hw_channel_dtr_try_complete(channel, rxdh); vxge_hw_channel_dtr_try_complete(channel, rxdh);
rxdp = (struct vxge_hw_ring_rxd_1 *)*rxdh; rxdp = *rxdh;
if (rxdp == NULL) { if (rxdp == NULL) {
status = VXGE_HW_INF_NO_MORE_COMPLETED_DESCRIPTORS; status = VXGE_HW_INF_NO_MORE_COMPLETED_DESCRIPTORS;
goto exit; goto exit;
@ -1565,7 +1565,7 @@ void vxge_hw_fifo_txdl_post(struct __vxge_hw_fifo *fifo, void *txdlh)
channel = &fifo->channel; channel = &fifo->channel;
txdl_priv = __vxge_hw_fifo_txdl_priv(fifo, txdlh); txdl_priv = __vxge_hw_fifo_txdl_priv(fifo, txdlh);
txdp_first = (struct vxge_hw_fifo_txd *)txdlh; txdp_first = txdlh;
txdp_last = (struct vxge_hw_fifo_txd *)txdlh + (txdl_priv->frags - 1); txdp_last = (struct vxge_hw_fifo_txd *)txdlh + (txdl_priv->frags - 1);
txdp_last->control_0 |= txdp_last->control_0 |=
@ -1631,7 +1631,7 @@ enum vxge_hw_status vxge_hw_fifo_txdl_next_completed(
vxge_hw_channel_dtr_try_complete(channel, txdlh); vxge_hw_channel_dtr_try_complete(channel, txdlh);
txdp = (struct vxge_hw_fifo_txd *)*txdlh; txdp = *txdlh;
if (txdp == NULL) { if (txdp == NULL) {
status = VXGE_HW_INF_NO_MORE_COMPLETED_DESCRIPTORS; status = VXGE_HW_INF_NO_MORE_COMPLETED_DESCRIPTORS;
goto exit; goto exit;

View file

@ -755,7 +755,7 @@ void cpc_tty_receive(pc300dev_t *pc300dev)
dsr_rx = cpc_readb(card->hw.scabase + DSR_RX(ch)); dsr_rx = cpc_readb(card->hw.scabase + DSR_RX(ch));
cpc_tty = (st_cpc_tty_area *)pc300dev->cpc_tty; cpc_tty = pc300dev->cpc_tty;
while (1) { while (1) {
rx_len = 0; rx_len = 0;

View file

@ -252,11 +252,11 @@ static void xemaclite_aligned_write(void *src_ptr, u32 *dest_ptr,
u16 *from_u16_ptr, *to_u16_ptr; u16 *from_u16_ptr, *to_u16_ptr;
to_u32_ptr = dest_ptr; to_u32_ptr = dest_ptr;
from_u16_ptr = (u16 *) src_ptr; from_u16_ptr = src_ptr;
align_buffer = 0; align_buffer = 0;
for (; length > 3; length -= 4) { for (; length > 3; length -= 4) {
to_u16_ptr = (u16 *) ((void *) &align_buffer); to_u16_ptr = (u16 *)&align_buffer;
*to_u16_ptr++ = *from_u16_ptr++; *to_u16_ptr++ = *from_u16_ptr++;
*to_u16_ptr++ = *from_u16_ptr++; *to_u16_ptr++ = *from_u16_ptr++;

View file

@ -442,19 +442,19 @@ static int __devinit yellowfin_init_one(struct pci_dev *pdev,
ring_space = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_cleardev; goto err_out_cleardev;
np->tx_ring = (struct yellowfin_desc *)ring_space; np->tx_ring = ring_space;
np->tx_ring_dma = ring_dma; np->tx_ring_dma = ring_dma;
ring_space = pci_alloc_consistent(pdev, RX_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pdev, RX_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_unmap_tx; goto err_out_unmap_tx;
np->rx_ring = (struct yellowfin_desc *)ring_space; np->rx_ring = ring_space;
np->rx_ring_dma = ring_dma; np->rx_ring_dma = ring_dma;
ring_space = pci_alloc_consistent(pdev, STATUS_TOTAL_SIZE, &ring_dma); ring_space = pci_alloc_consistent(pdev, STATUS_TOTAL_SIZE, &ring_dma);
if (!ring_space) if (!ring_space)
goto err_out_unmap_rx; goto err_out_unmap_rx;
np->tx_status = (struct tx_status_words *)ring_space; np->tx_status = ring_space;
np->tx_status_dma = ring_dma; np->tx_status_dma = ring_dma;
if (dev->mem_start) if (dev->mem_start)