Lines Matching defs:dev

123  * and dev->tx_timeout() should be called to fix the problem
964 if (device_may_wakeup(&tp->pdev->dev) &&
1544 tp->mdio_bus->parent = &tp->pdev->dev;
1559 dev_warn(&tp->pdev->dev, "mdiobus_reg failed (0x%x)\n", i);
1567 dev_warn(&tp->pdev->dev, "No PHY devices\n");
1840 netdev_info(tp->dev, "No firmware running\n");
1857 netdev_info(tp->dev, "No firmware running\n");
1872 if (!netif_carrier_ok(tp->dev)) {
1873 netif_info(tp, link, tp->dev, "Link is down\n");
1876 netdev_info(tp->dev, "Link is up at %d Mbps, %s duplex\n",
1884 netdev_info(tp->dev, "Flow control is %s for TX and %s for RX\n",
1891 netdev_info(tp->dev, "EEE is %s\n",
1897 tp->link_up = netif_carrier_ok(tp->dev);
1999 static void tg3_adjust_link(struct net_device *dev)
2003 struct tg3 *tp = netdev_priv(dev);
2096 phydev = phy_connect(tp->dev, phydev_name(phydev),
2099 dev_err(&tp->pdev->dev, "Could not attach to PHY\n");
2614 netif_carrier_off(tp->dev);
2621 netdev_warn(tp->dev,
2643 if (netif_running(tp->dev) && tp->link_up) {
2644 netif_carrier_off(tp->dev);
3656 netdev_err(tp->dev, "%s timed out, %s CPU\n",
3704 netdev_err(tp->dev,
3811 netdev_err(tp->dev, "%s fails to set RX CPU PC, is %08x "
3840 netdev_err(tp->dev, "Boot code not ready for service patches\n");
3846 netdev_warn(tp->dev,
3934 netdev_err(tp->dev,
3974 __tg3_set_one_mac_addr(tp, tp->dev->dev_addr, i);
3980 __tg3_set_one_mac_addr(tp, tp->dev->dev_addr, i);
3983 addr_high = (tp->dev->dev_addr[0] +
3984 tp->dev->dev_addr[1] +
3985 tp->dev->dev_addr[2] +
3986 tp->dev->dev_addr[3] +
3987 tp->dev->dev_addr[4] +
3988 tp->dev->dev_addr[5]) &
4014 netdev_err(tp->dev, "Transition to D0 failed\n");
4038 device_should_wake = device_may_wakeup(&tp->pdev->dev) &&
4712 netif_carrier_on(tp->dev);
4714 netif_carrier_off(tp->dev);
6155 static int tg3_get_ts_info(struct net_device *dev, struct ethtool_ts_info *info)
6157 struct tg3 *tp = netdev_priv(dev);
6281 netdev_warn(tp->dev,
6288 netdev_warn(tp->dev,
6446 netdev_err(tp->dev, "PCI channel ERROR!\n");
6466 netdev_err(tp->dev, "0x%08x: 0x%08x, 0x%08x, 0x%08x, 0x%08x\n",
6477 netdev_err(tp->dev,
6488 netdev_err(tp->dev,
6512 netdev_warn(tp->dev,
6545 txq = netdev_get_tx_queue(tp->dev, index);
6567 dma_unmap_single(&tp->pdev->dev, dma_unmap_addr(ri, mapping),
6585 dma_unmap_page(&tp->pdev->dev,
6647 dma_unmap_single(&tp->pdev->dev, dma_unmap_addr(ri, mapping), map_sz,
6712 mapping = dma_map_single(&tp->pdev->dev, data + TG3_RX_OFFSET(tp),
6714 if (unlikely(dma_mapping_error(&tp->pdev->dev, mapping))) {
6881 dma_unmap_single(&tp->pdev->dev, dma_addr, skb_size,
6904 skb = netdev_alloc_skb(tp->dev,
6910 dma_sync_single_for_cpu(&tp->pdev->dev, dma_addr, len,
6915 dma_sync_single_for_device(&tp->pdev->dev, dma_addr,
6924 if ((tp->dev->features & NETIF_F_RXCSUM) &&
6932 skb->protocol = eth_type_trans(skb, tp->dev);
6934 if (len > (tp->dev->mtu + ETH_HLEN) &&
7295 netdev_err(tp->dev, "FLOW Attention error. Resetting chip.\n");
7300 netdev_err(tp->dev, "MSI Status error. Resetting chip.\n");
7305 netdev_err(tp->dev, "DMA Status error. Resetting chip.\n");
7387 netif_napi_add(tp->dev, &tp->napi[0].napi, tg3_poll);
7389 netif_napi_add(tp->dev, &tp->napi[i].napi, tg3_poll_msix);
7402 netif_trans_update(tp->dev); /* prevent tx timeout */
7404 netif_carrier_off(tp->dev);
7405 netif_tx_disable(tp->dev);
7417 netif_tx_wake_all_queues(tp->dev);
7420 netif_carrier_on(tp->dev);
7624 static void tg3_poll_controller(struct net_device *dev)
7627 struct tg3 *tp = netdev_priv(dev);
7637 static void tg3_tx_timeout(struct net_device *dev, unsigned int txqueue)
7639 struct tg3 *tp = netdev_priv(dev);
7642 netdev_err(dev, "transmit timed out, resetting\n");
7766 dma_unmap_single(&tnapi->tp->pdev->dev, dma_unmap_addr(txb, mapping),
7781 dma_unmap_page(&tnapi->tp->pdev->dev,
7818 new_addr = dma_map_single(&tp->pdev->dev, new_skb->data,
7821 if (dma_mapping_error(&tp->pdev->dev, new_addr)) {
7883 segs = skb_gso_segment(skb, tp->dev->features &
7892 tg3_start_xmit(seg, tp->dev);
7902 static netdev_tx_t tg3_start_xmit(struct sk_buff *skb, struct net_device *dev)
7904 struct tg3 *tp = netdev_priv(dev);
7917 txq = netdev_get_tx_queue(dev, skb_get_queue_mapping(skb));
7934 netdev_err(dev,
8047 mapping = dma_map_single(&tp->pdev->dev, skb->data, len,
8049 if (dma_mapping_error(&tp->pdev->dev, mapping))
8081 mapping = skb_frag_dma_map(&tp->pdev->dev, frag, 0,
8087 if (dma_mapping_error(&tp->pdev->dev, mapping))
8284 static void tg3_set_loopback(struct net_device *dev, netdev_features_t features)
8286 struct tg3 *tp = netdev_priv(dev);
8294 netif_carrier_on(tp->dev);
8296 netdev_info(dev, "Internal MAC loopback mode enabled.\n");
8306 netdev_info(dev, "Internal MAC loopback mode disabled.\n");
8310 static netdev_features_t tg3_fix_features(struct net_device *dev,
8313 struct tg3 *tp = netdev_priv(dev);
8315 if (dev->mtu > ETH_DATA_LEN && tg3_flag(tp, 5780_CLASS))
8321 static int tg3_set_features(struct net_device *dev, netdev_features_t features)
8323 netdev_features_t changed = dev->features ^ features;
8325 if ((changed & NETIF_F_LOOPBACK) && netif_running(dev))
8326 tg3_set_loopback(dev, features);
8396 tp->dev->mtu > ETH_DATA_LEN)
8420 netdev_warn(tp->dev,
8455 netdev_warn(tp->dev,
8482 dma_free_coherent(&tp->pdev->dev, TG3_RX_STD_RING_BYTES(tp),
8487 dma_free_coherent(&tp->pdev->dev, TG3_RX_JMB_RING_BYTES(tp),
8501 tpr->rx_std = dma_alloc_coherent(&tp->pdev->dev,
8514 tpr->rx_jmb = dma_alloc_coherent(&tp->pdev->dev,
8559 netdev_tx_reset_queue(netdev_get_tx_queue(tp->dev, j));
8613 dma_free_coherent(&tp->pdev->dev, TG3_TX_RING_BYTES,
8641 tnapi->tx_ring = dma_alloc_coherent(&tp->pdev->dev,
8668 dma_free_coherent(&tp->pdev->dev,
8701 tnapi->rx_rcb = dma_alloc_coherent(&tp->pdev->dev,
8728 dma_free_coherent(&tp->pdev->dev, TG3_HW_STATUS_SIZE,
8743 dma_free_coherent(&tp->pdev->dev, sizeof(struct tg3_hw_stats),
8757 tp->hw_stats = dma_alloc_coherent(&tp->pdev->dev,
8767 tnapi->hw_status = dma_alloc_coherent(&tp->pdev->dev,
8848 dev_err(&tp->pdev->dev,
8862 dev_err(&tp->pdev->dev,
8917 dev_err(&tp->pdev->dev,
9371 static int tg3_set_mac_addr(struct net_device *dev, void *p)
9373 struct tg3 *tp = netdev_priv(dev);
9381 eth_hw_addr_set(dev, addr->sa_data);
9383 if (!netif_running(dev))
9401 __tg3_set_rx_mode(dev);
9743 static void __tg3_set_rx_mode(struct net_device *dev)
9745 struct tg3 *tp = netdev_priv(dev);
9759 if (dev->flags & IFF_PROMISC) {
9762 } else if (dev->flags & IFF_ALLMULTI) {
9765 } else if (netdev_mc_empty(dev)) {
9776 netdev_for_each_mc_addr(ha, dev) {
9790 if (netdev_uc_count(dev) > TG3_MAX_UCAST_ADDR(tp)) {
9792 } else if (!(dev->flags & IFF_PROMISC)) {
9797 netdev_for_each_uc_addr(ha, dev) {
10122 if (tp->dev->mtu <= ETH_DATA_LEN) {
10157 netdev_err(tp->dev, "%s cannot enable BUFMGR\n", __func__);
10246 tp->dev->mtu + ETH_HLEN + ETH_FCS_LEN + VLAN_HLEN);
10301 if (tp->dev->mtu <= ETH_DATA_LEN) {
10708 __tg3_set_rx_mode(tp->dev);
10815 static ssize_t tg3_show_temp(struct device *dev,
10819 struct tg3 *tp = dev_get_drvdata(dev);
10873 tp->hwmon_dev = hwmon_device_register_with_groups(&pdev->dev, "tg3",
10877 dev_err(&pdev->dev, "Cannot register hwmon device, aborting\n");
11169 netdev_err(tp->dev,
11176 dev_close(tp->dev);
11190 if (tp->pcierr_recovery || !netif_running(tp->dev) ||
11223 dev_close(tp->dev);
11243 name = tp->dev->name;
11248 "%s-txrx-%d", tp->dev->name, irq_num);
11251 "%s-tx-%d", tp->dev->name, irq_num);
11254 "%s-rx-%d", tp->dev->name, irq_num);
11257 "%s-%d", tp->dev->name, irq_num);
11279 struct net_device *dev = tp->dev;
11283 if (!netif_running(dev))
11300 IRQF_SHARED, dev->name, tnapi);
11380 netdev_warn(tp->dev, "No interrupt was generated using MSI. Switching "
11415 if (request_firmware(&tp->fw, tp->fw_needed, &tp->pdev->dev)) {
11416 netdev_err(tp->dev, "Failed to load firmware \"%s\"\n",
11430 netdev_err(tp->dev, "bogus length %d in \"%s\"\n",
11488 netdev_notice(tp->dev, "Requested %d MSI-X vectors, received %d\n",
11499 if (netif_set_real_num_rx_queues(tp->dev, tp->rxq_cnt)) {
11512 netif_set_real_num_tx_queues(tp->dev, tp->txq_cnt);
11524 netdev_warn(tp->dev,
11551 netif_set_real_num_tx_queues(tp->dev, 1);
11552 netif_set_real_num_rx_queues(tp->dev, 1);
11571 struct net_device *dev = tp->dev;
11655 netif_tx_start_all_queues(dev);
11661 if (dev->features & NETIF_F_LOOPBACK)
11662 tg3_set_loopback(dev, dev->features);
11718 static int tg3_open(struct net_device *dev)
11720 struct tg3 *tp = netdev_priv(dev);
11724 netdev_err(dev, "Failed to open device. PCI error recovery "
11733 netdev_warn(tp->dev, "EEE capability disabled\n");
11736 netdev_warn(tp->dev, "EEE capability restored\n");
11743 netdev_warn(tp->dev, "TSO capability disabled\n");
11746 netdev_notice(tp->dev, "TSO capability restored\n");
11775 static int tg3_close(struct net_device *dev)
11777 struct tg3 *tp = netdev_priv(dev);
11780 netdev_err(dev, "Failed to close device. PCI error recovery "
11987 static int tg3_get_regs_len(struct net_device *dev)
11992 static void tg3_get_regs(struct net_device *dev,
11995 struct tg3 *tp = netdev_priv(dev);
12011 static int tg3_get_eeprom_len(struct net_device *dev)
12013 struct tg3 *tp = netdev_priv(dev);
12018 static int tg3_get_eeprom(struct net_device *dev, struct ethtool_eeprom *eeprom, u8 *data)
12020 struct tg3 *tp = netdev_priv(dev);
12109 static int tg3_set_eeprom(struct net_device *dev, struct ethtool_eeprom *eeprom, u8 *data)
12111 struct tg3 *tp = netdev_priv(dev);
12165 static int tg3_get_link_ksettings(struct net_device *dev,
12168 struct tg3 *tp = netdev_priv(dev);
12217 if (netif_running(dev) && tp->link_up) {
12240 static int tg3_set_link_ksettings(struct net_device *dev,
12243 struct tg3 *tp = netdev_priv(dev);
12328 if (netif_running(dev))
12336 static void tg3_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info)
12338 struct tg3 *tp = netdev_priv(dev);
12345 static void tg3_get_wol(struct net_device *dev, struct ethtool_wolinfo *wol)
12347 struct tg3 *tp = netdev_priv(dev);
12349 if (tg3_flag(tp, WOL_CAP) && device_can_wakeup(&tp->pdev->dev))
12354 if (tg3_flag(tp, WOL_ENABLE) && device_can_wakeup(&tp->pdev->dev))
12359 static int tg3_set_wol(struct net_device *dev, struct ethtool_wolinfo *wol)
12361 struct tg3 *tp = netdev_priv(dev);
12362 struct device *dp = &tp->pdev->dev;
12380 static u32 tg3_get_msglevel(struct net_device *dev)
12382 struct tg3 *tp = netdev_priv(dev);
12386 static void tg3_set_msglevel(struct net_device *dev, u32 value)
12388 struct tg3 *tp = netdev_priv(dev);
12392 static int tg3_nway_reset(struct net_device *dev)
12394 struct tg3 *tp = netdev_priv(dev);
12397 if (!netif_running(dev))
12428 static void tg3_get_ringparam(struct net_device *dev,
12433 struct tg3 *tp = netdev_priv(dev);
12452 static int tg3_set_ringparam(struct net_device *dev,
12457 struct tg3 *tp = netdev_priv(dev);
12469 if (netif_running(dev)) {
12489 if (netif_running(dev)) {
12510 static void tg3_get_pauseparam(struct net_device *dev, struct ethtool_pauseparam *epause)
12512 struct tg3 *tp = netdev_priv(dev);
12527 static int tg3_set_pauseparam(struct net_device *dev, struct ethtool_pauseparam *epause)
12529 struct tg3 *tp = netdev_priv(dev);
12579 if (netif_running(dev)) {
12599 if (netif_running(dev)) {
12620 static int tg3_get_sset_count(struct net_device *dev, int sset)
12632 static int tg3_get_rxnfc(struct net_device *dev, struct ethtool_rxnfc *info,
12635 struct tg3 *tp = netdev_priv(dev);
12642 if (netif_running(tp->dev))
12657 static u32 tg3_get_rxfh_indir_size(struct net_device *dev)
12660 struct tg3 *tp = netdev_priv(dev);
12668 static int tg3_get_rxfh(struct net_device *dev, u32 *indir, u8 *key, u8 *hfunc)
12670 struct tg3 *tp = netdev_priv(dev);
12684 static int tg3_set_rxfh(struct net_device *dev, const u32 *indir, const u8 *key,
12687 struct tg3 *tp = netdev_priv(dev);
12703 if (!netif_running(dev) || !tg3_flag(tp, ENABLE_RSS))
12716 static void tg3_get_channels(struct net_device *dev,
12719 struct tg3 *tp = netdev_priv(dev);
12725 if (netif_running(dev)) {
12741 static int tg3_set_channels(struct net_device *dev,
12744 struct tg3 *tp = netdev_priv(dev);
12756 if (!netif_running(dev))
12768 static void tg3_get_strings(struct net_device *dev, u32 stringset, u8 *buf)
12783 static int tg3_set_phys_id(struct net_device *dev,
12786 struct tg3 *tp = netdev_priv(dev);
12815 static void tg3_get_ethtool_stats(struct net_device *dev,
12818 struct tg3 *tp = netdev_priv(dev);
13061 if (!netif_running(tp->dev))
13295 netdev_err(tp->dev,
13439 skb = netdev_alloc_skb(tp->dev, tx_len);
13444 memcpy(tx_data, tp->dev->dev_addr, ETH_ALEN);
13505 map = dma_map_single(&tp->pdev->dev, skb->data, tx_len, DMA_TO_DEVICE);
13506 if (dma_mapping_error(&tp->pdev->dev, map)) {
13604 dma_sync_single_for_cpu(&tp->pdev->dev, map, rx_len,
13641 if (!netif_running(tp->dev)) {
13744 static void tg3_self_test(struct net_device *dev, struct ethtool_test *etest,
13747 struct tg3 *tp = netdev_priv(dev);
13772 if (netif_running(dev)) {
13816 if (netif_running(dev)) {
13833 static int tg3_hwtstamp_set(struct net_device *dev, struct ifreq *ifr)
13835 struct tg3 *tp = netdev_priv(dev);
13904 if (netif_running(dev) && tp->rxptpctl)
13917 static int tg3_hwtstamp_get(struct net_device *dev, struct ifreq *ifr)
13919 struct tg3 *tp = netdev_priv(dev);
13978 static int tg3_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
13981 struct tg3 *tp = netdev_priv(dev);
14003 if (!netif_running(dev))
14020 if (!netif_running(dev))
14031 return tg3_hwtstamp_set(dev, ifr);
14034 return tg3_hwtstamp_get(dev, ifr);
14043 static int tg3_get_coalesce(struct net_device *dev,
14048 struct tg3 *tp = netdev_priv(dev);
14054 static int tg3_set_coalesce(struct net_device *dev,
14059 struct tg3 *tp = netdev_priv(dev);
14095 if (netif_running(dev)) {
14103 static int tg3_set_eee(struct net_device *dev, struct ethtool_eee *edata)
14105 struct tg3 *tp = netdev_priv(dev);
14108 netdev_warn(tp->dev, "Board does not support EEE!\n");
14113 netdev_warn(tp->dev,
14119 netdev_warn(tp->dev,
14130 if (netif_running(tp->dev)) {
14140 static int tg3_get_eee(struct net_device *dev, struct ethtool_eee *edata)
14142 struct tg3 *tp = netdev_priv(dev);
14145 netdev_warn(tp->dev,
14196 static void tg3_get_stats64(struct net_device *dev,
14199 struct tg3 *tp = netdev_priv(dev);
14212 static void tg3_set_rx_mode(struct net_device *dev)
14214 struct tg3 *tp = netdev_priv(dev);
14216 if (!netif_running(dev))
14220 __tg3_set_rx_mode(dev);
14224 static inline void tg3_set_mtu(struct net_device *dev, struct tg3 *tp,
14227 dev->mtu = new_mtu;
14231 netdev_update_features(dev);
14239 netdev_update_features(dev);
14245 static int tg3_change_mtu(struct net_device *dev, int new_mtu)
14247 struct tg3 *tp = netdev_priv(dev);
14251 if (!netif_running(dev)) {
14255 tg3_set_mtu(dev, tp, new_mtu);
14263 tg3_set_mtu(dev, tp, new_mtu);
15014 netdev_warn(tp->dev,
15166 device_set_wakeup_enable(&tp->pdev->dev, true);
15312 device_set_wakeup_enable(&tp->pdev->dev, true);
15355 device_set_wakeup_enable(&tp->pdev->dev,
15358 device_set_wakeup_capable(&tp->pdev->dev, false);
16458 dev_err(&tp->pdev->dev,
16705 if (tp->dev->mtu > ETH_DATA_LEN && !tg3_flag(tp, 5780_CLASS))
16873 dev_err(&tp->pdev->dev, "phy probe failed, err %d\n", err);
16956 if (!eth_platform_get_mac_address(&tp->pdev->dev, addr))
17262 buf = dma_alloc_coherent(&tp->pdev->dev, TEST_BUFFER_SIZE,
17367 dev_err(&tp->pdev->dev,
17376 dev_err(&tp->pdev->dev, "%s: Buffer read failed. "
17393 dev_err(&tp->pdev->dev,
17425 dma_free_coherent(&tp->pdev->dev, TEST_BUFFER_SIZE, buf, buf_dma);
17588 struct net_device *dev;
17599 dev_err(&pdev->dev, "Cannot enable PCI device, aborting\n");
17605 dev_err(&pdev->dev, "Cannot obtain PCI resources, aborting\n");
17611 dev = alloc_etherdev_mq(sizeof(*tp), TG3_IRQ_MAX_VECS);
17612 if (!dev) {
17617 SET_NETDEV_DEV(dev, &pdev->dev);
17619 tp = netdev_priv(dev);
17621 tp->dev = dev;
17673 dev_err(&pdev->dev, "Cannot map device registers, aborting\n");
17696 dev_err(&pdev->dev,
17706 dev->ethtool_ops = &tg3_ethtool_ops;
17707 dev->watchdog_timeo = TG3_TX_TIMEOUT;
17708 dev->netdev_ops = &tg3_netdev_ops;
17709 dev->irq = pdev->irq;
17713 dev_err(&pdev->dev,
17736 err = dma_set_mask(&pdev->dev, dma_mask);
17739 err = dma_set_coherent_mask(&pdev->dev,
17742 dev_err(&pdev->dev, "Unable to obtain 64 bit "
17749 err = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
17751 dev_err(&pdev->dev,
17790 dev->features |= features | NETIF_F_HW_VLAN_CTAG_TX |
17792 dev->vlan_features |= features;
17804 dev->hw_features |= features;
17805 dev->priv_flags |= IFF_UNICAST_FLT;
17808 dev->min_mtu = TG3_MIN_MTU;
17809 dev->max_mtu = TG3_MAX_MTU(tp);
17820 dev_err(&pdev->dev,
17824 eth_hw_addr_set(dev, addr);
17882 dev_err(&pdev->dev, "DMA engine test failed, aborting\n");
17888 pci_set_drvdata(pdev, dev);
17899 err = register_netdev(dev);
17901 dev_err(&pdev->dev, "Cannot register net device, aborting\n");
17908 &tp->pdev->dev);
17913 netdev_info(dev, "Tigon3 [partno(%s) rev %04x] (%s) MAC address %pM\n",
17917 dev->dev_addr);
17929 netdev_info(dev, "attached PHY is %s (%s Ethernet) "
17936 netdev_info(dev, "RXcsums[%d] LinkChgREG[%d] MIirq[%d] ASF[%d] TSOcap[%d]\n",
17937 (dev->features & NETIF_F_RXCSUM) != 0,
17942 netdev_info(dev, "dma_rwctrl[%08x] dma_mask[%d-bit]\n",
17964 free_netdev(dev);
17977 struct net_device *dev = pci_get_drvdata(pdev);
17979 if (dev) {
17980 struct tg3 *tp = netdev_priv(dev);
17993 unregister_netdev(dev);
18002 free_netdev(dev);
18011 struct net_device *dev = dev_get_drvdata(device);
18012 struct tg3 *tp = netdev_priv(dev);
18017 if (!netif_running(dev))
18030 netif_device_detach(dev);
18050 netif_device_attach(dev);
18067 struct net_device *dev = dev_get_drvdata(device);
18068 struct tg3 *tp = netdev_priv(dev);
18073 if (!netif_running(dev))
18076 netif_device_attach(dev);
18108 struct net_device *dev = pci_get_drvdata(pdev);
18109 struct tg3 *tp = netdev_priv(dev);
18115 netif_device_detach(dev);
18117 if (netif_running(dev))
18118 dev_close(dev);
18206 dev_err(&pdev->dev,