Lines Matching refs:card
150 * LCS free memory for card and channels.
153 lcs_free_card(struct lcs_card *card)
156 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
157 kfree(card);
161 * LCS alloc memory for card and channels
166 struct lcs_card *card;
171 card = kzalloc(sizeof(struct lcs_card), GFP_KERNEL | GFP_DMA);
172 if (card == NULL)
174 card->lan_type = LCS_FRAME_TYPE_AUTO;
175 card->pkt_seq = 0;
176 card->lancmd_timeout = LCS_LANCMD_TIMEOUT_DEFAULT;
178 rc = lcs_alloc_channel(&card->read);
181 lcs_free_card(card);
185 rc = lcs_alloc_channel(&card->write);
188 lcs_cleanup_channel(&card->read);
189 lcs_free_card(card);
194 INIT_LIST_HEAD(&card->ipm_list);
196 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
197 return card;
204 lcs_setup_read_ccws(struct lcs_card *card)
210 memset(card->read.ccws, 0, sizeof (struct ccw1) * (LCS_NUM_BUFFS + 1));
212 card->read.ccws[cnt].cmd_code = LCS_CCW_READ;
213 card->read.ccws[cnt].count = LCS_IOBUFFERSIZE;
214 card->read.ccws[cnt].flags =
220 card->read.ccws[cnt].cda =
221 (__u32)virt_to_phys(card->read.iob[cnt].data);
223 card->read.iob[cnt].data)->offset = LCS_ILLEGAL_OFFSET;
224 card->read.iob[cnt].callback = lcs_get_frames_cb;
225 card->read.iob[cnt].state = LCS_BUF_STATE_READY;
226 card->read.iob[cnt].count = LCS_IOBUFFERSIZE;
228 card->read.ccws[0].flags &= ~CCW_FLAG_PCI;
229 card->read.ccws[LCS_NUM_BUFFS - 1].flags &= ~CCW_FLAG_PCI;
230 card->read.ccws[LCS_NUM_BUFFS - 1].flags |= CCW_FLAG_SUSPEND;
232 card->read.ccws[LCS_NUM_BUFFS].cmd_code = LCS_CCW_TRANSFER;
233 card->read.ccws[LCS_NUM_BUFFS].cda =
234 (__u32)virt_to_phys(card->read.ccws);
236 card->read.state = LCS_CH_STATE_INIT;
238 card->read.io_idx = 0;
239 card->read.buf_idx = 0;
243 lcs_setup_read(struct lcs_card *card)
247 lcs_setup_read_ccws(card);
249 card->read.irq_tasklet.data = (unsigned long) &card->read;
250 card->read.irq_tasklet.func = lcs_tasklet;
252 init_waitqueue_head(&card->read.wait_q);
259 lcs_setup_write_ccws(struct lcs_card *card)
265 memset(card->write.ccws, 0, sizeof(struct ccw1) * (LCS_NUM_BUFFS + 1));
267 card->write.ccws[cnt].cmd_code = LCS_CCW_WRITE;
268 card->write.ccws[cnt].count = 0;
269 card->write.ccws[cnt].flags =
275 card->write.ccws[cnt].cda =
276 (__u32)virt_to_phys(card->write.iob[cnt].data);
279 card->write.ccws[LCS_NUM_BUFFS].cmd_code = LCS_CCW_TRANSFER;
280 card->write.ccws[LCS_NUM_BUFFS].cda =
281 (__u32)virt_to_phys(card->write.ccws);
283 card->read.state = LCS_CH_STATE_INIT;
285 card->write.io_idx = 0;
286 card->write.buf_idx = 0;
290 lcs_setup_write(struct lcs_card *card)
294 lcs_setup_write_ccws(card);
296 card->write.irq_tasklet.data = (unsigned long) &card->write;
297 card->write.irq_tasklet.func = lcs_tasklet;
299 init_waitqueue_head(&card->write.wait_q);
303 lcs_set_allowed_threads(struct lcs_card *card, unsigned long threads)
307 spin_lock_irqsave(&card->mask_lock, flags);
308 card->thread_allowed_mask = threads;
309 spin_unlock_irqrestore(&card->mask_lock, flags);
310 wake_up(&card->wait_q);
312 static int lcs_threads_running(struct lcs_card *card, unsigned long threads)
317 spin_lock_irqsave(&card->mask_lock, flags);
318 rc = (card->thread_running_mask & threads);
319 spin_unlock_irqrestore(&card->mask_lock, flags);
324 lcs_wait_for_threads(struct lcs_card *card, unsigned long threads)
326 return wait_event_interruptible(card->wait_q,
327 lcs_threads_running(card, threads) == 0);
330 static int lcs_set_thread_start_bit(struct lcs_card *card, unsigned long thread)
334 spin_lock_irqsave(&card->mask_lock, flags);
335 if ( !(card->thread_allowed_mask & thread) ||
336 (card->thread_start_mask & thread) ) {
337 spin_unlock_irqrestore(&card->mask_lock, flags);
340 card->thread_start_mask |= thread;
341 spin_unlock_irqrestore(&card->mask_lock, flags);
346 lcs_clear_thread_running_bit(struct lcs_card *card, unsigned long thread)
350 spin_lock_irqsave(&card->mask_lock, flags);
351 card->thread_running_mask &= ~thread;
352 spin_unlock_irqrestore(&card->mask_lock, flags);
353 wake_up(&card->wait_q);
356 static int __lcs_do_run_thread(struct lcs_card *card, unsigned long thread)
361 spin_lock_irqsave(&card->mask_lock, flags);
362 if (card->thread_start_mask & thread){
363 if ((card->thread_allowed_mask & thread) &&
364 !(card->thread_running_mask & thread)){
366 card->thread_start_mask &= ~thread;
367 card->thread_running_mask |= thread;
371 spin_unlock_irqrestore(&card->mask_lock, flags);
376 lcs_do_run_thread(struct lcs_card *card, unsigned long thread)
379 wait_event(card->wait_q,
380 (rc = __lcs_do_run_thread(card, thread)) >= 0);
385 lcs_do_start_thread(struct lcs_card *card, unsigned long thread)
390 spin_lock_irqsave(&card->mask_lock, flags);
392 (u8) card->thread_start_mask,
393 (u8) card->thread_allowed_mask,
394 (u8) card->thread_running_mask);
395 rc = (card->thread_start_mask & thread);
396 spin_unlock_irqrestore(&card->mask_lock, flags);
401 * Initialize channels,card and state machines.
404 lcs_setup_card(struct lcs_card *card)
407 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
409 lcs_setup_read(card);
410 lcs_setup_write(card);
412 card->state = DEV_STATE_DOWN;
413 card->tx_buffer = NULL;
414 card->tx_emitted = 0;
416 init_waitqueue_head(&card->wait_q);
417 spin_lock_init(&card->lock);
418 spin_lock_init(&card->ipm_lock);
419 spin_lock_init(&card->mask_lock);
421 INIT_LIST_HEAD(&card->ipm_list);
423 INIT_LIST_HEAD(&card->lancmd_waiters);
426 static void lcs_clear_multicast_list(struct lcs_card *card)
434 spin_lock_irqsave(&card->ipm_lock, flags);
435 while (!list_empty(&card->ipm_list)){
436 ipm = list_entry(card->ipm_list.next,
440 spin_unlock_irqrestore(&card->ipm_lock, flags);
441 lcs_send_delipm(card, ipm);
442 spin_lock_irqsave(&card->ipm_lock, flags);
446 spin_unlock_irqrestore(&card->ipm_lock, flags);
451 * Cleanup channels,card and state machines.
454 lcs_cleanup_card(struct lcs_card *card)
458 LCS_DBF_HEX(2,setup,&card,sizeof(void*));
460 if (card->dev != NULL)
461 free_netdev(card->dev);
463 lcs_cleanup_channel(&card->write);
464 lcs_cleanup_channel(&card->read);
548 lcs_start_channels(struct lcs_card *card)
554 rc = lcs_start_channel(&card->read);
558 rc = lcs_start_channel(&card->write);
560 lcs_stop_channel(&card->read);
568 lcs_stop_channels(struct lcs_card *card)
571 lcs_stop_channel(&card->read);
572 lcs_stop_channel(&card->write);
731 lcs_get_lancmd(struct lcs_card *card, int count)
738 wait_event(card->write.wait_q,
739 ((buffer = lcs_get_buffer(&card->write)) != NULL));
788 lcs_notify_lancmd_waiters(struct lcs_card *card, struct lcs_cmd *cmd)
794 spin_lock(&card->lock);
795 list_for_each_safe(l, n, &card->lancmd_waiters) {
801 reply->callback(card, cmd);
809 spin_unlock(&card->lock);
823 spin_lock_irqsave(&reply->card->lock, flags);
825 &reply->card->lancmd_waiters,list) {
829 spin_unlock_irqrestore(&reply->card->lock, flags);
837 spin_unlock_irqrestore(&reply->card->lock, flags);
841 lcs_send_lancmd(struct lcs_card *card, struct lcs_buffer *buffer,
852 cmd->sequence_no = card->sequence_no++;
857 reply->card = card;
858 spin_lock_irqsave(&card->lock, flags);
859 list_add_tail(&reply->list, &card->lancmd_waiters);
860 spin_unlock_irqrestore(&card->lock, flags);
863 rc = lcs_ready_buffer(&card->write, buffer);
867 mod_timer(&reply->timer, jiffies + HZ * card->lancmd_timeout);
880 lcs_send_startup(struct lcs_card *card, __u8 initiator)
886 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
891 return lcs_send_lancmd(card, buffer, NULL);
898 lcs_send_shutdown(struct lcs_card *card)
904 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
908 return lcs_send_lancmd(card, buffer, NULL);
915 __lcs_lanstat_cb(struct lcs_card *card, struct lcs_cmd *cmd)
918 memcpy(card->mac, cmd->cmd.lcs_lanstat_cmd.mac_addr, LCS_MAC_LENGTH);
922 lcs_send_lanstat(struct lcs_card *card)
928 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
933 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
934 cmd->cmd.lcs_std_cmd.portno = card->portno;
935 return lcs_send_lancmd(card, buffer, __lcs_lanstat_cb);
942 lcs_send_stoplan(struct lcs_card *card, __u8 initiator)
948 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
952 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
953 cmd->cmd.lcs_std_cmd.portno = card->portno;
954 return lcs_send_lancmd(card, buffer, NULL);
961 __lcs_send_startlan_cb(struct lcs_card *card, struct lcs_cmd *cmd)
964 card->lan_type = cmd->cmd.lcs_std_cmd.lan_type;
965 card->portno = cmd->cmd.lcs_std_cmd.portno;
969 lcs_send_startlan(struct lcs_card *card, __u8 initiator)
975 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
979 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
980 cmd->cmd.lcs_std_cmd.portno = card->portno;
981 return lcs_send_lancmd(card, buffer, __lcs_send_startlan_cb);
989 lcs_send_setipm(struct lcs_card *card,struct lcs_ipm_list *ipm_list)
995 buffer = lcs_get_lancmd(card, LCS_MULTICAST_CMD_SIZE);
999 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
1000 cmd->cmd.lcs_qipassist.portno = card->portno;
1006 return lcs_send_lancmd(card, buffer, NULL);
1013 lcs_send_delipm(struct lcs_card *card,struct lcs_ipm_list *ipm_list)
1019 buffer = lcs_get_lancmd(card, LCS_MULTICAST_CMD_SIZE);
1023 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
1024 cmd->cmd.lcs_qipassist.portno = card->portno;
1030 return lcs_send_lancmd(card, buffer, NULL);
1037 __lcs_check_multicast_cb(struct lcs_card *card, struct lcs_cmd *cmd)
1040 card->ip_assists_supported =
1042 card->ip_assists_enabled =
1047 lcs_check_multicast_support(struct lcs_card *card)
1055 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
1059 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
1060 cmd->cmd.lcs_qipassist.portno = card->portno;
1063 rc = lcs_send_lancmd(card, buffer, __lcs_check_multicast_cb);
1068 if (card->ip_assists_supported & LCS_IPASS_MULTICAST_SUPPORT)
1074 * set or del multicast address on LCS card
1077 lcs_fix_multicast_list(struct lcs_card *card)
1086 spin_lock_irqsave(&card->ipm_lock, flags);
1088 list_for_each_entry_safe(ipm, tmp, &card->ipm_list, list){
1094 spin_unlock_irqrestore(&card->ipm_lock, flags);
1095 rc = lcs_send_setipm(card, ipm);
1096 spin_lock_irqsave(&card->ipm_lock, flags);
1107 list_add_tail(&ipm->list, &card->ipm_list);
1112 spin_unlock_irqrestore(&card->ipm_lock, flags);
1113 lcs_send_delipm(card, ipm);
1114 spin_lock_irqsave(&card->ipm_lock, flags);
1123 list_move_tail(&ipm->list, &card->ipm_list);
1125 spin_unlock_irqrestore(&card->ipm_lock, flags);
1141 static void lcs_remove_mc_addresses(struct lcs_card *card,
1151 spin_lock_irqsave(&card->ipm_lock, flags);
1152 list_for_each(l, &card->ipm_list) {
1156 lcs_get_mac_for_ipm(im4->multiaddr, buf, card->dev);
1165 spin_unlock_irqrestore(&card->ipm_lock, flags);
1168 static struct lcs_ipm_list *lcs_check_addr_entry(struct lcs_card *card,
1177 spin_lock_irqsave(&card->ipm_lock, flags);
1178 list_for_each(l, &card->ipm_list) {
1187 spin_unlock_irqrestore(&card->ipm_lock, flags);
1191 static void lcs_set_mc_addresses(struct lcs_card *card,
1203 lcs_get_mac_for_ipm(im4->multiaddr, buf, card->dev);
1204 ipm = lcs_check_addr_entry(card, im4, buf);
1216 spin_lock_irqsave(&card->ipm_lock, flags);
1218 list_add(&ipm->list, &card->ipm_list);
1219 spin_unlock_irqrestore(&card->ipm_lock, flags);
1226 struct lcs_card *card;
1229 card = (struct lcs_card *) data;
1231 if (!lcs_do_run_thread(card, LCS_SET_MC_THREAD))
1235 in4_dev = in_dev_get(card->dev);
1239 lcs_remove_mc_addresses(card,in4_dev);
1240 lcs_set_mc_addresses(card, in4_dev);
1244 netif_carrier_off(card->dev);
1245 netif_tx_disable(card->dev);
1246 wait_event(card->write.wait_q,
1247 (card->write.state != LCS_CH_STATE_RUNNING));
1248 lcs_fix_multicast_list(card);
1249 if (card->state == DEV_STATE_UP) {
1250 netif_carrier_on(card->dev);
1251 netif_wake_queue(card->dev);
1254 lcs_clear_thread_running_bit(card, LCS_SET_MC_THREAD);
1267 struct lcs_card *card;
1270 card = (struct lcs_card *) dev->ml_priv;
1272 if (!lcs_set_thread_start_bit(card, LCS_SET_MC_THREAD))
1273 schedule_work(&card->kernel_thread_starter);
1347 lcs_schedule_recovery(struct lcs_card *card)
1350 if (!lcs_set_thread_start_bit(card, LCS_RECOVERY_THREAD))
1351 schedule_work(&card->kernel_thread_starter);
1360 struct lcs_card *card;
1368 card = CARD_FROM_DEV(cdev);
1369 if (card->read.ccwdev == cdev)
1370 channel = &card->read;
1372 channel = &card->write;
1394 lcs_schedule_recovery(card);
1395 wake_up(&card->wait_q);
1479 __lcs_emit_txbuffer(struct lcs_card *card)
1482 *(__u16 *)(card->tx_buffer->data + card->tx_buffer->count) = 0;
1483 card->tx_buffer->count += 2;
1484 lcs_ready_buffer(&card->write, card->tx_buffer);
1485 card->tx_buffer = NULL;
1486 card->tx_emitted++;
1495 struct lcs_card *card;
1500 card = container_of(channel, struct lcs_card, write);
1501 if (netif_queue_stopped(card->dev) && netif_carrier_ok(card->dev))
1502 netif_wake_queue(card->dev);
1503 spin_lock(&card->lock);
1504 card->tx_emitted--;
1505 if (card->tx_emitted <= 0 && card->tx_buffer != NULL)
1510 __lcs_emit_txbuffer(card);
1511 spin_unlock(&card->lock);
1517 static netdev_tx_t __lcs_start_xmit(struct lcs_card *card, struct sk_buff *skb,
1525 card->stats.tx_dropped++;
1526 card->stats.tx_errors++;
1529 if (card->state != DEV_STATE_UP) {
1531 card->stats.tx_dropped++;
1532 card->stats.tx_errors++;
1533 card->stats.tx_carrier_errors++;
1540 netif_stop_queue(card->dev);
1541 spin_lock(&card->lock);
1542 if (card->tx_buffer != NULL &&
1543 card->tx_buffer->count + sizeof(struct lcs_header) +
1546 __lcs_emit_txbuffer(card);
1547 if (card->tx_buffer == NULL) {
1549 card->tx_buffer = lcs_get_buffer(&card->write);
1550 if (card->tx_buffer == NULL) {
1551 card->stats.tx_dropped++;
1555 card->tx_buffer->callback = lcs_txbuffer_cb;
1556 card->tx_buffer->count = 0;
1559 (card->tx_buffer->data + card->tx_buffer->count);
1560 card->tx_buffer->count += skb->len + sizeof(struct lcs_header);
1561 header->offset = card->tx_buffer->count;
1562 header->type = card->lan_type;
1563 header->slot = card->portno;
1565 spin_unlock(&card->lock);
1566 card->stats.tx_bytes += skb->len;
1567 card->stats.tx_packets++;
1569 netif_wake_queue(card->dev);
1570 spin_lock(&card->lock);
1571 if (card->tx_emitted <= 0 && card->tx_buffer != NULL)
1573 __lcs_emit_txbuffer(card);
1575 spin_unlock(&card->lock);
1581 struct lcs_card *card;
1585 card = (struct lcs_card *) dev->ml_priv;
1586 rc = __lcs_start_xmit(card, skb, dev);
1594 lcs_startlan_auto(struct lcs_card *card)
1599 card->lan_type = LCS_FRAME_TYPE_ENET;
1600 rc = lcs_send_startlan(card, LCS_INITIATOR_TCPIP);
1608 lcs_startlan(struct lcs_card *card)
1614 if (card->portno != LCS_INVALID_PORT_NO) {
1615 if (card->lan_type == LCS_FRAME_TYPE_AUTO)
1616 rc = lcs_startlan_auto(card);
1618 rc = lcs_send_startlan(card, LCS_INITIATOR_TCPIP);
1621 card->portno = i;
1622 if (card->lan_type != LCS_FRAME_TYPE_AUTO)
1623 rc = lcs_send_startlan(card,
1627 rc = lcs_startlan_auto(card);
1633 return lcs_send_lanstat(card);
1642 lcs_detect(struct lcs_card *card)
1647 /* start/reset card */
1648 if (card->dev)
1649 netif_stop_queue(card->dev);
1650 rc = lcs_stop_channels(card);
1652 rc = lcs_start_channels(card);
1654 rc = lcs_send_startup(card, LCS_INITIATOR_TCPIP);
1656 rc = lcs_startlan(card);
1660 card->state = DEV_STATE_UP;
1662 card->state = DEV_STATE_DOWN;
1663 card->write.state = LCS_CH_STATE_INIT;
1664 card->read.state = LCS_CH_STATE_INIT;
1670 * LCS Stop card
1673 lcs_stopcard(struct lcs_card *card)
1679 if (card->read.state != LCS_CH_STATE_STOPPED &&
1680 card->write.state != LCS_CH_STATE_STOPPED &&
1681 card->read.state != LCS_CH_STATE_ERROR &&
1682 card->write.state != LCS_CH_STATE_ERROR &&
1683 card->state == DEV_STATE_UP) {
1684 lcs_clear_multicast_list(card);
1685 rc = lcs_send_stoplan(card,LCS_INITIATOR_TCPIP);
1686 rc = lcs_send_shutdown(card);
1688 rc = lcs_stop_channels(card);
1689 card->state = DEV_STATE_DOWN;
1700 struct lcs_card *card = container_of(work, struct lcs_card, kernel_thread_starter);
1702 if (lcs_do_start_thread(card, LCS_RECOVERY_THREAD))
1703 kthread_run(lcs_recovery, card, "lcs_recover");
1705 if (lcs_do_start_thread(card, LCS_SET_MC_THREAD))
1706 kthread_run(lcs_register_mc_addresses, card, "regipm");
1714 lcs_get_control(struct lcs_card *card, struct lcs_cmd *cmd)
1721 lcs_schedule_recovery(card);
1724 if (card->dev) {
1726 card->dev->name);
1727 netif_carrier_off(card->dev);
1735 lcs_notify_lancmd_waiters(card, cmd);
1742 lcs_get_skb(struct lcs_card *card, char *skb_data, unsigned int skb_len)
1747 if (card->dev == NULL ||
1748 card->state != DEV_STATE_UP)
1749 /* The card isn't up. Ignore the packet. */
1754 dev_err(&card->dev->dev,
1756 card->dev->name);
1757 card->stats.rx_dropped++;
1761 skb->protocol = card->lan_type_trans(skb, card->dev);
1762 card->stats.rx_bytes += skb_len;
1763 card->stats.rx_packets++;
1765 *((__u32 *)skb->cb) = ++card->pkt_seq;
1775 struct lcs_card *card;
1785 card = container_of(channel, struct lcs_card, read);
1792 card->stats.rx_length_errors++;
1793 card->stats.rx_errors++;
1797 lcs_get_control(card, (struct lcs_cmd *) lcs_hdr);
1799 lcs_get_skb(card, (char *)(lcs_hdr + 1),
1803 dev_info_once(&card->dev->dev,
1811 lcs_ready_buffer(&card->read, buffer);
1820 struct lcs_card *card;
1823 card = (struct lcs_card *) dev->ml_priv;
1824 return &card->stats;
1834 struct lcs_card *card;
1838 card = (struct lcs_card *) dev->ml_priv;
1842 wait_event(card->write.wait_q,
1843 (card->write.state != LCS_CH_STATE_RUNNING));
1844 rc = lcs_stopcard(card);
1846 dev_err(&card->dev->dev,
1858 struct lcs_card *card;
1862 card = (struct lcs_card *) dev->ml_priv;
1864 rc = lcs_detect(card);
1872 card->state = DEV_STATE_UP;
1883 struct lcs_card *card;
1885 card = dev_get_drvdata(dev);
1887 if (!card)
1890 return sysfs_emit(buf, "%d\n", card->portno);
1899 struct lcs_card *card;
1903 card = dev_get_drvdata(dev);
1905 if (!card)
1912 card->portno = value;
1913 if (card->dev)
1914 card->dev->dev_port = card->portno;
1926 "OSA LCS card",
1949 struct lcs_card *card;
1951 card = dev_get_drvdata(dev);
1953 return card ? sysfs_emit(buf, "%u\n", card->lancmd_timeout) : 0;
1959 struct lcs_card *card;
1963 card = dev_get_drvdata(dev);
1965 if (!card)
1972 card->lancmd_timeout = value;
1984 struct lcs_card *card = dev_get_drvdata(dev);
1988 if (!card)
1990 if (card->state != DEV_STATE_UP)
1994 lcs_schedule_recovery(card);
2025 struct lcs_card *card;
2031 card = lcs_alloc_card();
2032 if (!card) {
2037 dev_set_drvdata(&ccwgdev->dev, card);
2040 card->gdev = ccwgdev;
2041 INIT_WORK(&card->kernel_thread_starter, lcs_start_kernel_thread);
2042 card->thread_start_mask = 0;
2043 card->thread_allowed_mask = 0;
2044 card->thread_running_mask = 0;
2053 struct lcs_card *card;
2056 card = dev_get_drvdata(&ccwgdev->dev);
2057 if (card->dev->reg_state != NETREG_UNINITIALIZED)
2059 SET_NETDEV_DEV(card->dev, &ccwgdev->dev);
2060 return register_netdev(card->dev);
2084 struct lcs_card *card;
2089 card = dev_get_drvdata(&ccwgdev->dev);
2090 if (!card)
2094 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2095 card->read.ccwdev = ccwgdev->cdev[0];
2096 card->write.ccwdev = ccwgdev->cdev[1];
2098 recover_state = card->state;
2099 rc = ccw_device_set_online(card->read.ccwdev);
2102 rc = ccw_device_set_online(card->write.ccwdev);
2108 lcs_setup_card(card);
2109 rc = lcs_detect(card);
2115 lcs_stopcard(card);
2118 if (card->dev) {
2120 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2123 switch (card->lan_type) {
2125 card->lan_type_trans = eth_type_trans;
2135 card->dev = dev;
2136 card->dev->ml_priv = card;
2137 card->dev->netdev_ops = &lcs_netdev_ops;
2138 card->dev->dev_port = card->portno;
2139 eth_hw_addr_set(card->dev, card->mac);
2141 if (!lcs_check_multicast_support(card))
2142 card->dev->netdev_ops = &lcs_mc_netdev_ops;
2145 lcs_set_allowed_threads(card,0xffffffff);
2147 lcs_set_multicast_list(card->dev);
2148 card->dev->flags |= IFF_UP;
2149 netif_carrier_on(card->dev);
2150 netif_wake_queue(card->dev);
2151 card->state = DEV_STATE_UP;
2153 lcs_stopcard(card);
2160 pr_info("LCS device %s %s IPv6 support\n", card->dev->name,
2161 (card->ip_assists_supported & LCS_IPASS_IPV6_SUPPORT) ?
2164 pr_info("LCS device %s %s Multicast support\n", card->dev->name,
2165 (card->ip_assists_supported & LCS_IPASS_MULTICAST_SUPPORT) ?
2170 ccw_device_set_offline(card->write.ccwdev);
2172 ccw_device_set_offline(card->read.ccwdev);
2183 struct lcs_card *card;
2188 card = dev_get_drvdata(&ccwgdev->dev);
2189 if (!card)
2192 lcs_set_allowed_threads(card, 0);
2193 if (lcs_wait_for_threads(card, LCS_SET_MC_THREAD))
2196 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2197 recover_state = card->state;
2199 ret = lcs_stop_device(card->dev);
2200 ret2 = ccw_device_set_offline(card->read.ccwdev);
2201 ret3 = ccw_device_set_offline(card->write.ccwdev);
2207 card->state = DEV_STATE_RECOVER;
2224 struct lcs_card *card;
2228 card = (struct lcs_card *) ptr;
2231 if (!lcs_do_run_thread(card, LCS_RECOVERY_THREAD))
2234 gdev = card->gdev;
2241 card->dev->name);
2244 card->dev->name);
2245 lcs_clear_thread_running_bit(card, LCS_RECOVERY_THREAD);
2250 * lcs_remove_device, free buffers and card
2255 struct lcs_card *card;
2257 card = dev_get_drvdata(&ccwgdev->dev);
2258 if (!card)
2262 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2266 if (card->dev)
2267 unregister_netdev(card->dev);
2268 lcs_cleanup_card(card);
2269 lcs_free_card(card);