Lines Matching defs:ndev

137 		netif_alert(qdev, probe, qdev->ndev, "bad Semaphore mask!.\n");
177 netif_alert(qdev, probe, qdev->ndev,
186 netif_alert(qdev, probe, qdev->ndev,
230 netif_err(qdev, ifup, qdev->ndev, "Couldn't map DMA area.\n");
240 netif_err(qdev, ifup, qdev->ndev,
322 netif_crit(qdev, ifup, qdev->ndev,
413 if (qdev->ndev->features & NETIF_F_HW_VLAN_CTAG_RX)
440 netif_crit(qdev, ifup, qdev->ndev,
460 netif_printk(qdev, ifup, KERN_DEBUG, qdev->ndev,
465 netif_printk(qdev, ifup, KERN_DEBUG, qdev->ndev,
475 netif_err(qdev, ifup, qdev->ndev,
482 netif_err(qdev, link, qdev->ndev, "Link is up.\n");
483 netif_carrier_on(qdev->ndev);
489 netif_err(qdev, link, qdev->ndev, "Link is down.\n");
490 netif_carrier_off(qdev->ndev);
600 netif_err(qdev, ifup, qdev->ndev,
707 netif_err(qdev, ifup, qdev->ndev, "Invalid flash signature.\n");
715 netif_err(qdev, ifup, qdev->ndev,
768 netif_err(qdev, ifup, qdev->ndev,
778 netif_err(qdev, ifup, qdev->ndev, "Invalid flash.\n");
789 qdev->ndev->addr_len);
793 qdev->ndev->addr_len);
796 netif_err(qdev, ifup, qdev->ndev, "Invalid MAC address.\n");
801 memcpy(qdev->ndev->dev_addr,
803 qdev->ndev->addr_len);
830 netif_err(qdev, ifup, qdev->ndev,
841 netif_err(qdev, ifup, qdev->ndev, "Invalid flash.\n");
851 memcpy(qdev->ndev->dev_addr,
853 qdev->ndev->addr_len);
959 netif_info(qdev, link, qdev->ndev,
963 netif_crit(qdev, link, qdev->ndev,
969 netif_info(qdev, link, qdev->ndev, "Got xgmac semaphore!.\n");
1099 netif_err(qdev, drv, qdev->ndev,
1111 netif_err(qdev, drv, qdev->ndev,
1149 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1155 netif_err(qdev, ifup, qdev->ndev,
1184 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1204 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1209 qdev->ndev,
1213 netdev_alloc_skb(qdev->ndev,
1225 netif_err(qdev, ifup, qdev->ndev,
1250 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1285 qdev->ndev,
1295 netif_printk(qdev, tx_done, KERN_DEBUG, qdev->ndev,
1321 netif_printk(qdev, tx_queued, KERN_DEBUG, qdev->ndev,
1331 netif_err(qdev, tx_queued, qdev->ndev,
1378 netif_err(qdev, tx_queued, qdev->ndev,
1406 netif_err(qdev, tx_queued, qdev->ndev,
1478 if (qdev->ndev->features & NETIF_F_HW_VLAN_CTAG_RX)
1508 napi->dev = qdev->ndev;
1512 netif_err(qdev, drv, qdev->ndev,
1545 struct net_device *ndev = qdev->ndev;
1552 skb = netdev_alloc_skb(ndev, length);
1574 if (skb->len > ndev->mtu + hlen) {
1575 netif_err(qdev, drv, qdev->ndev,
1581 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1593 skb->protocol = eth_type_trans(skb, ndev);
1596 if ((ndev->features & NETIF_F_RXCSUM) &&
1600 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1612 qdev->ndev,
1638 struct net_device *ndev = qdev->ndev;
1645 new_skb = netdev_alloc_skb(qdev->ndev, length + NET_IP_ALIGN);
1671 if (skb->len > ndev->mtu + ETH_HLEN) {
1679 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1689 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1694 skb->protocol = eth_type_trans(skb, ndev);
1700 if ((ndev->features & NETIF_F_RXCSUM) &&
1704 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1715 qdev->ndev,
1765 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1785 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1792 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1820 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1837 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1846 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1862 skb = netdev_alloc_skb(qdev->ndev, length);
1864 netif_printk(qdev, probe, KERN_DEBUG, qdev->ndev,
1874 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1918 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1930 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1956 struct net_device *ndev = qdev->ndev;
1963 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1979 if (skb->len > ndev->mtu + ETH_HLEN) {
1994 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev, "%s Multicast.\n",
2004 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
2008 skb->protocol = eth_type_trans(skb, ndev);
2014 if ((ndev->features & NETIF_F_RXCSUM) &&
2018 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
2028 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
2052 (qdev->ndev->features & NETIF_F_HW_VLAN_CTAG_RX)) ?
2117 netif_warn(qdev, tx_done, qdev->ndev,
2121 netif_warn(qdev, tx_done, qdev->ndev,
2125 netif_warn(qdev, tx_done, qdev->ndev,
2129 netif_warn(qdev, tx_done, qdev->ndev,
2164 netif_err(qdev, rx_err, qdev->ndev,
2170 netdev_err(qdev->ndev, "Multiple CAM hits lookup occurred.\n");
2171 netdev_err(qdev->ndev, "This event shouldn't occur.\n");
2176 netdev_err(qdev->ndev, "Soft ECC error detected.\n");
2181 netdev_err(qdev->ndev, "PCI error occurred when reading "
2188 netif_err(qdev, drv, qdev->ndev, "Unexpected event %d.\n",
2206 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
2219 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
2231 if (__netif_subqueue_stopped(qdev->ndev, tx_ring->wq_id)) {
2237 netif_wake_subqueue(qdev->ndev, tx_ring->wq_id);
2253 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
2271 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
2295 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
2308 netif_printk(qdev, intr, KERN_DEBUG, qdev->ndev,
2320 netif_printk(qdev, intr, KERN_DEBUG, qdev->ndev,
2333 static void qlge_vlan_mode(struct net_device *ndev, netdev_features_t features)
2335 struct ql_adapter *qdev = netdev_priv(ndev);
2349 static int qlge_update_hw_vlan_features(struct net_device *ndev,
2352 struct ql_adapter *qdev = netdev_priv(ndev);
2357 netif_err(qdev, link, qdev->ndev,
2363 ndev->features = features;
2367 netif_err(qdev, link, qdev->ndev,
2374 static netdev_features_t qlge_fix_features(struct net_device *ndev,
2380 err = qlge_update_hw_vlan_features(ndev, features);
2387 static int qlge_set_features(struct net_device *ndev,
2390 netdev_features_t changed = ndev->features ^ features;
2393 qlge_vlan_mode(ndev, features);
2406 netif_err(qdev, ifup, qdev->ndev,
2411 static int qlge_vlan_rx_add_vid(struct net_device *ndev, __be16 proto, u16 vid)
2413 struct ql_adapter *qdev = netdev_priv(ndev);
2437 netif_err(qdev, ifup, qdev->ndev,
2442 static int qlge_vlan_rx_kill_vid(struct net_device *ndev, __be16 proto, u16 vid)
2444 struct ql_adapter *qdev = netdev_priv(ndev);
2498 netif_printk(qdev, intr, KERN_DEBUG, qdev->ndev,
2512 netdev_err(qdev->ndev, "Got fatal error, STS = %x.\n", var);
2514 netdev_err(qdev->ndev, "Resetting chip. "
2528 netif_err(qdev, intr, qdev->ndev,
2544 netif_info(qdev, intr, qdev->ndev,
2627 static netdev_tx_t qlge_send(struct sk_buff *skb, struct net_device *ndev)
2631 struct ql_adapter *qdev = netdev_priv(ndev);
2642 netif_info(qdev, tx_queued, qdev->ndev,
2645 netif_stop_subqueue(ndev, tx_ring->wq_id);
2664 netif_printk(qdev, tx_queued, KERN_DEBUG, qdev->ndev,
2679 netif_err(qdev, tx_queued, qdev->ndev,
2691 netif_printk(qdev, tx_queued, KERN_DEBUG, qdev->ndev,
2698 netif_stop_subqueue(ndev, tx_ring->wq_id);
2704 netif_wake_subqueue(qdev->ndev, tx_ring->wq_id);
2734 netif_err(qdev, ifup, qdev->ndev,
2743 netif_err(qdev, ifup, qdev->ndev,
2809 netif_err(qdev, ifup, qdev->ndev, "tx_ring alloc failed.\n");
2855 netif_err(qdev, ifup, qdev->ndev,
2983 netif_err(qdev, ifup, qdev->ndev, "rx_ring alloc failed.\n");
2996 netif_err(qdev, ifup, qdev->ndev,
3022 netif_err(qdev, ifup, qdev->ndev,
3060 netif_err(qdev, ifdown, qdev->ndev,
3094 netif_err(qdev, ifup, qdev->ndev,
3102 netif_err(qdev, ifup, qdev->ndev,
3232 netif_napi_add(qdev->ndev, &rx_ring->napi, ql_napi_poll_msix,
3238 netif_printk(qdev, ifup, KERN_DEBUG, qdev->ndev,
3244 netif_err(qdev, ifup, qdev->ndev, "Failed to load CQICB.\n");
3290 netif_err(qdev, ifup, qdev->ndev, "Failed to load tx_ring.\n");
3338 netif_warn(qdev, ifup, qdev->ndev,
3344 netif_info(qdev, ifup, qdev->ndev,
3355 netif_info(qdev, ifup, qdev->ndev,
3361 netif_printk(qdev, ifup, KERN_DEBUG, qdev->ndev,
3478 qdev->ndev->name, i);
3485 qdev->ndev->name, i);
3510 sprintf(intr_context->name, "%s-single_irq", qdev->ndev->name);
3560 netif_err(qdev, ifup, qdev->ndev,
3566 netif_printk(qdev, ifup, KERN_DEBUG, qdev->ndev,
3568 netif_printk(qdev, ifup, KERN_DEBUG, qdev->ndev,
3570 netif_printk(qdev, ifup, KERN_DEBUG, qdev->ndev,
3573 netif_printk(qdev, ifup, KERN_DEBUG, qdev->ndev,
3585 netif_err(qdev, ifup, qdev->ndev,
3598 netif_err(qdev, ifup, qdev->ndev, "Failed to get the interrupts!!!\n");
3635 netif_err(qdev, ifup, qdev->ndev, "Failed to load RICB.\n");
3652 netif_err(qdev, ifup, qdev->ndev,
3678 netif_err(qdev, ifup, qdev->ndev,
3686 netif_err(qdev, ifup, qdev->ndev,
3693 netif_err(qdev, ifup, qdev->ndev,
3704 netif_err(qdev, ifup, qdev->ndev,
3713 netif_err(qdev, ifup, qdev->ndev,
3732 netif_err(qdev, ifup, qdev->ndev, "Failed to init mac address.\n");
3738 netif_err(qdev, ifup, qdev->ndev, "Failed to init routing table.\n");
3759 if (qdev->ndev->features & NETIF_F_HW_VLAN_CTAG_RX) {
3807 netif_err(qdev, ifup, qdev->ndev,
3819 netif_err(qdev, ifup, qdev->ndev, "Failed to start RSS.\n");
3828 netif_err(qdev, ifup, qdev->ndev,
3837 netif_err(qdev, ifup, qdev->ndev, "Failed to start port.\n");
3842 netif_err(qdev, ifup, qdev->ndev,
3864 netif_err(qdev, ifup, qdev->ndev, "Failed to clear routing bits.\n");
3893 netif_err(qdev, ifdown, qdev->ndev,
3903 static void ql_display_dev_info(struct net_device *ndev)
3905 struct ql_adapter *qdev = netdev_priv(ndev);
3907 netif_info(qdev, probe, qdev->ndev,
3916 netif_info(qdev, probe, qdev->ndev,
3917 "MAC address %pM\n", ndev->dev_addr);
3934 netif_err(qdev, ifdown, qdev->ndev,
3943 netif_err(qdev, ifdown, qdev->ndev,
3945 qdev->ndev->name);
3948 netif_info(qdev, drv, qdev->ndev,
3950 qdev->ndev->name);
3958 netif_err(qdev, drv, qdev->ndev,
3961 wol, qdev->ndev->name);
4006 netif_err(qdev, ifdown, qdev->ndev, "reset(func #%d) FAILED!\n",
4019 netif_info(qdev, ifup, qdev->ndev, "Unable to initialize adapter.\n");
4033 qlge_set_multicast_list(qdev->ndev);
4040 netif_tx_start_all_queues(qdev->ndev);
4059 netif_err(qdev, ifup, qdev->ndev, "Unable to allocate memory.\n");
4066 static int qlge_close(struct net_device *ndev)
4068 struct ql_adapter *qdev = netdev_priv(ndev);
4075 netif_err(qdev, drv, qdev->ndev, "EEH fatal did unload.\n");
4097 unsigned int lbq_buf_len = (qdev->ndev->mtu > 1500) ?
4174 static int qlge_open(struct net_device *ndev)
4177 struct ql_adapter *qdev = netdev_priv(ndev);
4212 netif_err(qdev, ifup, qdev->ndev,
4218 netif_err(qdev, ifup, qdev->ndev,
4229 lbq_buf_len = (qdev->ndev->mtu > 1500) ?
4245 netif_alert(qdev, ifup, qdev->ndev,
4248 dev_close(qdev->ndev);
4252 static int qlge_change_mtu(struct net_device *ndev, int new_mtu)
4254 struct ql_adapter *qdev = netdev_priv(ndev);
4257 if (ndev->mtu == 1500 && new_mtu == 9000) {
4258 netif_err(qdev, ifup, qdev->ndev, "Changing to jumbo MTU.\n");
4259 } else if (ndev->mtu == 9000 && new_mtu == 1500) {
4260 netif_err(qdev, ifup, qdev->ndev, "Changing to normal MTU.\n");
4267 ndev->mtu = new_mtu;
4269 if (!netif_running(qdev->ndev)) {
4275 netif_err(qdev, ifup, qdev->ndev,
4283 *ndev)
4285 struct ql_adapter *qdev = netdev_priv(ndev);
4300 ndev->stats.rx_packets = pkts;
4301 ndev->stats.rx_bytes = bytes;
4302 ndev->stats.rx_dropped = dropped;
4303 ndev->stats.rx_errors = errors;
4304 ndev->stats.multicast = mcast;
4313 ndev->stats.tx_packets = pkts;
4314 ndev->stats.tx_bytes = bytes;
4315 ndev->stats.tx_errors = errors;
4316 return &ndev->stats;
4319 static void qlge_set_multicast_list(struct net_device *ndev)
4321 struct ql_adapter *qdev = netdev_priv(ndev);
4332 if (ndev->flags & IFF_PROMISC) {
4336 netif_err(qdev, hw, qdev->ndev,
4346 netif_err(qdev, hw, qdev->ndev,
4358 if ((ndev->flags & IFF_ALLMULTI) ||
4359 (netdev_mc_count(ndev) > MAX_MULTICAST_ENTRIES)) {
4363 netif_err(qdev, hw, qdev->ndev,
4373 netif_err(qdev, hw, qdev->ndev,
4381 if (!netdev_mc_empty(ndev)) {
4386 netdev_for_each_mc_addr(ha, ndev) {
4389 netif_err(qdev, hw, qdev->ndev,
4399 netif_err(qdev, hw, qdev->ndev,
4409 static int qlge_set_mac_address(struct net_device *ndev, void *p)
4411 struct ql_adapter *qdev = netdev_priv(ndev);
4417 memcpy(ndev->dev_addr, addr->sa_data, ndev->addr_len);
4419 memcpy(qdev->current_mac_addr, ndev->dev_addr, ndev->addr_len);
4424 status = ql_set_mac_addr_reg(qdev, (u8 *) ndev->dev_addr,
4427 netif_err(qdev, hw, qdev->ndev, "Failed to load MAC address.\n");
4432 static void qlge_tx_timeout(struct net_device *ndev)
4434 struct ql_adapter *qdev = netdev_priv(ndev);
4455 qlge_set_multicast_list(qdev->ndev);
4460 netif_alert(qdev, ifup, qdev->ndev,
4464 dev_close(qdev->ndev);
4548 struct net_device *ndev = pci_get_drvdata(pdev);
4549 struct ql_adapter *qdev = netdev_priv(ndev);
4564 static int ql_init_device(struct pci_dev *pdev, struct net_device *ndev,
4567 struct ql_adapter *qdev = netdev_priv(ndev);
4577 qdev->ndev = ndev;
4579 pci_set_drvdata(pdev, ndev);
4659 memcpy(qdev->current_mac_addr, ndev->dev_addr, ndev->addr_len);
4674 qdev->workqueue = create_singlethread_workqueue(ndev->name);
4720 netif_err(qdev, ifup, qdev->ndev, "EEH STS = 0x%.08x.\n", var);
4730 struct net_device *ndev = NULL;
4735 ndev = alloc_etherdev_mq(sizeof(struct ql_adapter),
4737 if (!ndev)
4740 err = ql_init_device(pdev, ndev, cards_found);
4742 free_netdev(ndev);
4746 qdev = netdev_priv(ndev);
4747 SET_NETDEV_DEV(ndev, &pdev->dev);
4748 ndev->hw_features = NETIF_F_SG |
4756 ndev->features = ndev->hw_features;
4757 ndev->vlan_features = ndev->hw_features;
4759 ndev->vlan_features &= ~(NETIF_F_HW_VLAN_CTAG_FILTER |
4764 ndev->features |= NETIF_F_HIGHDMA;
4769 ndev->tx_queue_len = qdev->tx_ring_size;
4770 ndev->irq = pdev->irq;
4772 ndev->netdev_ops = &qlge_netdev_ops;
4773 ndev->ethtool_ops = &qlge_ethtool_ops;
4774 ndev->watchdog_timeo = 10 * HZ;
4776 err = register_netdev(ndev);
4781 free_netdev(ndev);
4793 ql_display_dev_info(ndev);
4799 netdev_tx_t ql_lb_send(struct sk_buff *skb, struct net_device *ndev)
4801 return qlge_send(skb, ndev);
4811 struct net_device *ndev = pci_get_drvdata(pdev);
4812 struct ql_adapter *qdev = netdev_priv(ndev);
4815 unregister_netdev(ndev);
4818 free_netdev(ndev);
4822 static void ql_eeh_close(struct net_device *ndev)
4825 struct ql_adapter *qdev = netdev_priv(ndev);
4827 if (netif_carrier_ok(ndev)) {
4828 netif_carrier_off(ndev);
4829 netif_stop_queue(ndev);
4852 struct net_device *ndev = pci_get_drvdata(pdev);
4853 struct ql_adapter *qdev = netdev_priv(ndev);
4859 netif_device_detach(ndev);
4860 if (netif_running(ndev))
4861 ql_eeh_close(ndev);
4867 ql_eeh_close(ndev);
4884 struct net_device *ndev = pci_get_drvdata(pdev);
4885 struct ql_adapter *qdev = netdev_priv(ndev);
4891 netif_err(qdev, ifup, qdev->ndev,
4898 netif_err(qdev, drv, qdev->ndev, "reset FAILED!\n");
4908 struct net_device *ndev = pci_get_drvdata(pdev);
4909 struct ql_adapter *qdev = netdev_priv(ndev);
4912 if (netif_running(ndev)) {
4913 err = qlge_open(ndev);
4915 netif_err(qdev, ifup, qdev->ndev,
4920 netif_err(qdev, ifup, qdev->ndev,
4924 netif_device_attach(ndev);
4935 struct net_device *ndev = pci_get_drvdata(pdev);
4936 struct ql_adapter *qdev = netdev_priv(ndev);
4939 netif_device_detach(ndev);
4942 if (netif_running(ndev)) {
4963 struct net_device *ndev = pci_get_drvdata(pdev);
4964 struct ql_adapter *qdev = netdev_priv(ndev);
4971 netif_err(qdev, ifup, qdev->ndev, "Cannot enable PCI device from suspend\n");
4979 if (netif_running(ndev)) {
4986 netif_device_attach(ndev);