📄 ioc3-eth.c
字号:
int i, phy; spin_lock_irq(&ip->ioc3_lock); phy = -1; for (i = 0; i < 32; i++) { word = mii_read(ioc3, i, 2); if ((word != 0xffff) && (word != 0x0000)) { phy = i; break; /* Found a PHY */ } } if (phy == -1) { spin_unlock_irq(&ip->ioc3_lock); return -ENODEV; } ip->phy = phy; /* Autonegotiate 100mbit and fullduplex. */ mii0 = mii_read(ioc3, ip->phy, 0); mii_write(ioc3, ip->phy, 0, mii0 | 0x3100); ip->negtimer.function = &negotiate; ip->negtimer.data = (unsigned long) dev; mod_timer(&ip->negtimer, jiffies); /* Run it now */ spin_unlock_irq(&ip->ioc3_lock); return 0;}static inline voidioc3_clean_rx_ring(struct ioc3_private *ip){ struct sk_buff *skb; int i; for (i = ip->rx_ci; i & 15; i++) { ip->rx_skbs[ip->rx_pi] = ip->rx_skbs[ip->rx_ci]; ip->rxr[ip->rx_pi++] = ip->rxr[ip->rx_ci++]; } ip->rx_pi &= 511; ip->rx_ci &= 511; for (i = ip->rx_ci; i != ip->rx_pi; i = (i+1) & 511) { struct ioc3_erxbuf *rxb; skb = ip->rx_skbs[i]; rxb = (struct ioc3_erxbuf *) (skb->data - RX_OFFSET); rxb->w0 = 0; }}static inline voidioc3_clean_tx_ring(struct ioc3_private *ip){ struct sk_buff *skb; int i; for (i=0; i < 128; i++) { skb = ip->tx_skbs[i]; if (skb) { ip->tx_skbs[i] = NULL; dev_kfree_skb_any(skb); } ip->txr[i].cmd = 0; } ip->tx_pi = 0; ip->tx_ci = 0;}static voidioc3_free_rings(struct ioc3_private *ip){ struct sk_buff *skb; int rx_entry, n_entry; if (ip->txr) { ioc3_clean_tx_ring(ip); free_pages((unsigned long)ip->txr, 2); ip->txr = NULL; } if (ip->rxr) { n_entry = ip->rx_ci; rx_entry = ip->rx_pi; while (n_entry != rx_entry) { skb = ip->rx_skbs[n_entry]; if (skb) dev_kfree_skb_any(skb); n_entry = (n_entry + 1) & 511; } free_page((unsigned long)ip->rxr); ip->rxr = NULL; }}static voidioc3_alloc_rings(struct net_device *dev, struct ioc3_private *ip, struct ioc3 *ioc3){ struct ioc3_erxbuf *rxb; unsigned long *rxr; int i; if (ip->rxr == NULL) { /* Allocate and initialize rx ring. 4kb = 512 entries */ ip->rxr = (unsigned long *) get_free_page(GFP_KERNEL|GFP_ATOMIC); rxr = (unsigned long *) ip->rxr; /* Now the rx buffers. The RX ring may be larger but we only allocate 16 buffers for now. Need to tune this for performance and memory later. */ for (i = 0; i < RX_BUFFS; i++) { struct sk_buff *skb; skb = ioc3_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); if (!skb) { show_free_areas(); continue; } ip->rx_skbs[i] = skb; skb->dev = dev; /* Because we reserve afterwards. */ skb_put(skb, (1664 + RX_OFFSET)); rxb = (struct ioc3_erxbuf *) skb->data; rxr[i] = (0xa5UL << 56) | ((unsigned long) rxb & TO_PHYS_MASK); skb_reserve(skb, RX_OFFSET); } ip->rx_ci = 0; ip->rx_pi = RX_BUFFS; } if (ip->txr == NULL) { /* Allocate and initialize tx rings. 16kb = 128 bufs. */ ip->txr = (struct ioc3_etxd *)__get_free_pages(GFP_KERNEL|GFP_ATOMIC, 2); ip->tx_pi = 0; ip->tx_ci = 0; }}static voidioc3_init_rings(struct net_device *dev, struct ioc3_private *ip, struct ioc3 *ioc3){ unsigned long ring; ioc3_free_rings(ip); ioc3_alloc_rings(dev, ip, ioc3); ioc3_clean_rx_ring(ip); ioc3_clean_tx_ring(ip); /* Now the rx ring base, consume & produce registers. */ ring = (0xa5UL << 56) | ((unsigned long)ip->rxr & TO_PHYS_MASK); ioc3->erbr_h = ring >> 32; ioc3->erbr_l = ring & 0xffffffff; ioc3->ercir = (ip->rx_ci << 3); ioc3->erpir = (ip->rx_pi << 3) | ERPIR_ARM; ring = (0xa5UL << 56) | ((unsigned long)ip->txr & TO_PHYS_MASK); ip->txqlen = 0; /* nothing queued */ /* Now the tx ring base, consume & produce registers. */ ioc3->etbr_h = ring >> 32; ioc3->etbr_l = ring & 0xffffffff; ioc3->etpir = (ip->tx_pi << 7); ioc3->etcir = (ip->tx_ci << 7); ioc3->etcir; /* Flush */}static inline voidioc3_ssram_disc(struct ioc3_private *ip){ struct ioc3 *ioc3 = ip->regs; volatile u32 *ssram0 = &ioc3->ssram[0x0000]; volatile u32 *ssram1 = &ioc3->ssram[0x4000]; unsigned int pattern = 0x5555; /* Assume the larger size SSRAM and enable parity checking */ ioc3->emcr |= (EMCR_BUFSIZ | EMCR_RAMPAR); *ssram0 = pattern; *ssram1 = ~pattern & IOC3_SSRAM_DM; if ((*ssram0 & IOC3_SSRAM_DM) != pattern || (*ssram1 & IOC3_SSRAM_DM) != (~pattern & IOC3_SSRAM_DM)) { /* set ssram size to 64 KB */ ip->emcr = EMCR_RAMPAR; ioc3->emcr &= ~EMCR_BUFSIZ; } else { ip->emcr = EMCR_BUFSIZ | EMCR_RAMPAR; }}static void ioc3_init(struct net_device *dev){ struct ioc3_private *ip = dev->priv; struct ioc3 *ioc3 = ip->regs; ioc3->emcr = EMCR_RST; /* Reset */ ioc3->emcr; /* flush WB */ udelay(4); /* Give it time ... */ ioc3->emcr = 0; ioc3->emcr; /* Misc registers */ ioc3->erbar = 0; ioc3->etcsr = (17<<ETCSR_IPGR2_SHIFT) | (11<<ETCSR_IPGR1_SHIFT) | 21; ioc3->etcdc; /* Clear on read */ ioc3->ercsr = 15; /* RX low watermark */ ioc3->ertr = 0; /* Interrupt immediately */ ioc3->emar_h = (dev->dev_addr[5] << 8) | dev->dev_addr[4]; ioc3->emar_l = (dev->dev_addr[3] << 24) | (dev->dev_addr[2] << 16) | (dev->dev_addr[1] << 8) | dev->dev_addr[0]; ioc3->ehar_h = ip->ehar_h; ioc3->ehar_l = ip->ehar_l; ioc3->ersr = 42; /* XXX should be random */ ioc3_init_rings(dev, ip, ioc3); ip->emcr |= ((RX_OFFSET / 2) << EMCR_RXOFF_SHIFT) | EMCR_TXDMAEN | EMCR_TXEN | EMCR_RXDMAEN | EMCR_RXEN; ioc3->emcr = ip->emcr; ioc3->eier = EISR_RXTIMERINT | EISR_RXOFLO | EISR_RXBUFOFLO | EISR_RXMEMERR | EISR_RXPARERR | EISR_TXBUFUFLO | EISR_TXEXPLICIT | EISR_TXMEMERR; ioc3->eier;}static inline void ioc3_stop(struct net_device *dev){ struct ioc3_private *ip = dev->priv; struct ioc3 *ioc3 = ip->regs; ioc3->emcr = 0; /* Shutup */ ioc3->eier = 0; /* Disable interrupts */ ioc3->eier; /* Flush */}static intioc3_open(struct net_device *dev){ struct ioc3_private *ip; if (request_irq(dev->irq, ioc3_interrupt, 0, ioc3_str, dev)) { printk(KERN_ERR "%s: Can't get irq %d\n", dev->name, dev->irq); return -EAGAIN; } ip = (struct ioc3_private *) dev->priv; ip->ehar_h = 0; ip->ehar_l = 0; ioc3_init(dev); netif_start_queue(dev); MOD_INC_USE_COUNT; return 0;}static intioc3_close(struct net_device *dev){ struct ioc3_private *ip = dev->priv; del_timer(&ip->negtimer); netif_stop_queue(dev); ioc3_stop(dev); /* Flush */ free_irq(dev->irq, dev); ioc3_free_rings(ip); MOD_DEC_USE_COUNT; return 0;}static int ioc3_pci_init(struct pci_dev *pdev){ u16 mii0, mii_status, mii2, mii3, mii4; struct net_device *dev = NULL; // XXX struct ioc3_private *ip; struct ioc3 *ioc3; unsigned long ioc3_base, ioc3_size; u32 vendor, model, rev; int phy; dev = init_etherdev(0, sizeof(struct ioc3_private)); if (!dev) return -ENOMEM; ip = dev->priv; memset(ip, 0, sizeof(*ip)); /* * This probably needs to be register_netdevice, or call * init_etherdev so that it calls register_netdevice. Quick * hack for now. */ netif_device_attach(dev); dev->irq = pdev->irq; ioc3_base = pdev->resource[0].start; ioc3_size = pdev->resource[0].end - ioc3_base; ioc3 = (struct ioc3 *) ioremap(ioc3_base, ioc3_size); ip->regs = ioc3; spin_lock_init(&ip->ioc3_lock); ioc3_stop(dev); ip->emcr = 0; ioc3_init(dev); init_timer(&ip->negtimer); ioc3_mii_init(dev, ip, ioc3); phy = ip->phy; if (phy == -1) { printk(KERN_CRIT"%s: Didn't find a PHY, goodbye.\n", dev->name); ioc3_stop(dev); free_irq(dev->irq, dev); ioc3_free_rings(ip); return -ENODEV; } mii0 = mii_read(ioc3, phy, 0); mii_status = mii_read(ioc3, phy, 1); mii2 = mii_read(ioc3, phy, 2); mii3 = mii_read(ioc3, phy, 3); mii4 = mii_read(ioc3, phy, 4); vendor = (mii2 << 12) | (mii3 >> 4); model = (mii3 >> 4) & 0x3f; rev = mii3 & 0xf; printk(KERN_INFO"Using PHY %d, vendor 0x%x, model %d, rev %d.\n", phy, vendor, model, rev); printk(KERN_INFO "%s: MII transceiver found at MDIO address " "%d, config %4.4x status %4.4x.\n", dev->name, phy, mii0, mii_status); ioc3_ssram_disc(ip); printk("IOC3 SSRAM has %d kbyte.\n", ip->emcr & EMCR_BUFSIZ ? 128 : 64); ioc3_get_eaddr(dev, ioc3); /* The IOC3-specific entries in the device structure. */ dev->open = ioc3_open; dev->hard_start_xmit = ioc3_start_xmit; dev->tx_timeout = ioc3_timeout; dev->watchdog_timeo = 5 * HZ; dev->stop = ioc3_close; dev->get_stats = ioc3_get_stats; dev->do_ioctl = ioc3_ioctl; dev->set_multicast_list = ioc3_set_multicast_list; return 0;}static int __init ioc3_probe(void){ static int called = 0; int cards = 0; if (called) return -ENODEV; called = 1; if (pci_present()) { struct pci_dev *pdev = NULL; while ((pdev = pci_find_device(PCI_VENDOR_ID_SGI, PCI_DEVICE_ID_SGI_IOC3, pdev))) { if (ioc3_pci_init(pdev)) return -ENOMEM; cards++; } } return cards ? -ENODEV : 0;}static void __exit ioc3_cleanup_module(void){ /* Later, when we really support modules. */}static intioc3_start_xmit(struct sk_buff *skb, struct net_device *dev){ unsigned long data; struct ioc3_private *ip = dev->priv; struct ioc3 *ioc3 = ip->regs; unsigned int len; struct ioc3_etxd *desc; int produce; spin_lock_irq(&ip->ioc3_lock); data = (unsigned long) skb->data; len = skb->len; produce = ip->tx_pi; desc = &ip->txr[produce]; if (len <= 104) { /* Short packet, let's copy it directly into the ring. */ memcpy(desc->data, skb->data, skb->len); if (len < ETH_ZLEN) { /* Very short packet, pad with zeros at the end. */ memset(desc->data + len, 0, ETH_ZLEN - len); len = ETH_ZLEN; } desc->cmd = len | ETXD_INTWHENDONE | ETXD_D0V; desc->bufcnt = len; } else if ((data ^ (data + len)) & 0x4000) { unsigned long b2, s1, s2; b2 = (data | 0x3fffUL) + 1UL; s1 = b2 - data; s2 = data + len - b2; desc->cmd = len | ETXD_INTWHENDONE | ETXD_B1V | ETXD_B2V; desc->bufcnt = (s1 << ETXD_B1CNT_SHIFT) | (s2 << ETXD_B2CNT_SHIFT); desc->p1 = (0xa5UL << 56) | (data & TO_PHYS_MASK); desc->p2 = (0xa5UL << 56) | (data & TO_PHYS_MASK); } else { /* Normal sized packet that doesn't cross a page boundary. */ desc->cmd = len | ETXD_INTWHENDONE | ETXD_B1V; desc->bufcnt = len << ETXD_B1CNT_SHIFT; desc->p1 = (0xa5UL << 56) | (data & TO_PHYS_MASK); } BARRIER(); dev->trans_start = jiffies; ip->tx_skbs[produce] = skb; /* Remember skb */ produce = (produce + 1) & 127; ip->tx_pi = produce; ioc3->etpir = produce << 7; /* Fire ... */ ip->txqlen++; if (ip->txqlen > 127) netif_stop_queue(dev); spin_unlock_irq(&ip->ioc3_lock); return 0;}static void ioc3_timeout(struct net_device *dev){ struct ioc3_private *ip = dev->priv; struct ioc3 *ioc3 = ip->regs; printk(KERN_ERR "%s: transmit timed out, resetting\n", dev->name); ioc3_stop(dev); ioc3_init(dev); ioc3_mii_init(dev, ip, ioc3); dev->trans_start = jiffies; netif_wake_queue(dev);}/* * Given a multicast ethernet address, this routine calculates the * address's bit index in the logical address filter mask */#define CRC_MASK 0xEDB88320static inline unsigned intioc3_hash(const unsigned char *addr){ unsigned int temp = 0; unsigned char byte; unsigned int crc; int bits, len; len = ETH_ALEN; for (crc = ~0; --len >= 0; addr++) { byte = *addr; for (bits = 8; --bits >= 0; ) { if ((byte ^ crc) & 1) crc = (crc >> 1) ^ CRC_MASK; else crc >>= 1; byte >>= 1; } } crc &= 0x3f; /* bit reverse lowest 6 bits for hash index */ for (bits = 6; --bits >= 0; ) { temp <<= 1; temp |= (crc & 0x1); crc >>= 1; } return temp;}/* Provide ioctl() calls to examine the MII xcvr state. */static int ioc3_ioctl(struct net_device *dev, struct ifreq *rq, int cmd){ struct ioc3_private *ip = (struct ioc3_private *) dev->priv; u16 *data = (u16 *)&rq->ifr_data; struct ioc3 *ioc3 = ip->regs; int phy = ip->phy; switch (cmd) { case SIOCGMIIPHY: /* Get the address of the PHY in use. */ if (phy == -1) return -ENODEV; data[0] = phy; return 0; case SIOCGMIIREG: /* Read any PHY register. */ spin_lock_irq(&ip->ioc3_lock); data[3] = mii_read(ioc3, data[0], data[1]); spin_unlock_irq(&ip->ioc3_lock); return 0; case SIOCSMIIREG: /* Write any PHY register. */ if (!capable(CAP_NET_ADMIN)) return -EPERM; spin_lock_irq(&ip->ioc3_lock); mii_write(ioc3, data[0], data[1], data[2]); spin_unlock_irq(&ip->ioc3_lock); return 0; default: return -EOPNOTSUPP; } return -EOPNOTSUPP;}static void ioc3_set_multicast_list(struct net_device *dev){ struct dev_mc_list *dmi = dev->mc_list; struct ioc3_private *ip = dev->priv; struct ioc3 *ioc3 = ip->regs; char *addr = dmi->dmi_addr; u64 ehar = 0; int i; if (dev->flags & IFF_PROMISC) { /* Set promiscuous. */ /* Unconditionally log net taps. */ printk(KERN_INFO "%s: Promiscuous mode enabled.\n", dev->name); ip->emcr |= EMCR_PROMISC; ioc3->emcr = ip->emcr; ioc3->emcr; } else { ip->emcr &= ~EMCR_PROMISC; ioc3->emcr = ip->emcr; /* Clear promiscuous. */ ioc3->emcr; if ((dev->flags & IFF_ALLMULTI) || (dev->mc_count > 64)) { /* Too many for hashing to make sense or we want all multicast packets anyway, so skip computing all the hashes and just accept all packets. */ ip->ehar_h = 0xffffffff; ip->ehar_l = 0xffffffff; } else { for (i = 0; i < dev->mc_count; i++) { dmi = dmi->next; if (!(*addr & 1)) continue; ehar |= (1UL << ioc3_hash(addr)); } ip->ehar_h = ehar >> 32; ip->ehar_l = ehar & 0xffffffff; } ioc3->ehar_h = ip->ehar_h; ioc3->ehar_l = ip->ehar_l; }}#ifdef MODULEMODULE_AUTHOR("Ralf Baechle <ralf@oss.sgi.com>");MODULE_DESCRIPTION("SGI IOC3 Ethernet driver");#endif /* MODULE */module_init(ioc3_probe);module_exit(ioc3_cleanup_module);
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -