Lines Matching refs:card

155  * LCS free memory for card and channels.
158 lcs_free_card(struct lcs_card *card)
161 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
162 kfree(card);
166 * LCS alloc memory for card and channels
171 struct lcs_card *card;
176 card = kzalloc(sizeof(struct lcs_card), GFP_KERNEL | GFP_DMA);
177 if (card == NULL)
179 card->lan_type = LCS_FRAME_TYPE_AUTO;
180 card->pkt_seq = 0;
181 card->lancmd_timeout = LCS_LANCMD_TIMEOUT_DEFAULT;
183 rc = lcs_alloc_channel(&card->read);
186 lcs_free_card(card);
190 rc = lcs_alloc_channel(&card->write);
193 lcs_cleanup_channel(&card->read);
194 lcs_free_card(card);
199 INIT_LIST_HEAD(&card->ipm_list);
201 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
202 return card;
209 lcs_setup_read_ccws(struct lcs_card *card)
215 memset(card->read.ccws, 0, sizeof (struct ccw1) * (LCS_NUM_BUFFS + 1));
217 card->read.ccws[cnt].cmd_code = LCS_CCW_READ;
218 card->read.ccws[cnt].count = LCS_IOBUFFERSIZE;
219 card->read.ccws[cnt].flags =
225 card->read.ccws[cnt].cda =
226 (__u32) __pa(card->read.iob[cnt].data);
228 card->read.iob[cnt].data)->offset = LCS_ILLEGAL_OFFSET;
229 card->read.iob[cnt].callback = lcs_get_frames_cb;
230 card->read.iob[cnt].state = LCS_BUF_STATE_READY;
231 card->read.iob[cnt].count = LCS_IOBUFFERSIZE;
233 card->read.ccws[0].flags &= ~CCW_FLAG_PCI;
234 card->read.ccws[LCS_NUM_BUFFS - 1].flags &= ~CCW_FLAG_PCI;
235 card->read.ccws[LCS_NUM_BUFFS - 1].flags |= CCW_FLAG_SUSPEND;
237 card->read.ccws[LCS_NUM_BUFFS].cmd_code = LCS_CCW_TRANSFER;
238 card->read.ccws[LCS_NUM_BUFFS].cda =
239 (__u32) __pa(card->read.ccws);
241 card->read.state = LCS_CH_STATE_INIT;
243 card->read.io_idx = 0;
244 card->read.buf_idx = 0;
248 lcs_setup_read(struct lcs_card *card)
252 lcs_setup_read_ccws(card);
254 card->read.irq_tasklet.data = (unsigned long) &card->read;
255 card->read.irq_tasklet.func = lcs_tasklet;
257 init_waitqueue_head(&card->read.wait_q);
264 lcs_setup_write_ccws(struct lcs_card *card)
270 memset(card->write.ccws, 0, sizeof(struct ccw1) * (LCS_NUM_BUFFS + 1));
272 card->write.ccws[cnt].cmd_code = LCS_CCW_WRITE;
273 card->write.ccws[cnt].count = 0;
274 card->write.ccws[cnt].flags =
280 card->write.ccws[cnt].cda =
281 (__u32) __pa(card->write.iob[cnt].data);
284 card->write.ccws[LCS_NUM_BUFFS].cmd_code = LCS_CCW_TRANSFER;
285 card->write.ccws[LCS_NUM_BUFFS].cda =
286 (__u32) __pa(card->write.ccws);
288 card->read.state = LCS_CH_STATE_INIT;
290 card->write.io_idx = 0;
291 card->write.buf_idx = 0;
295 lcs_setup_write(struct lcs_card *card)
299 lcs_setup_write_ccws(card);
301 card->write.irq_tasklet.data = (unsigned long) &card->write;
302 card->write.irq_tasklet.func = lcs_tasklet;
304 init_waitqueue_head(&card->write.wait_q);
308 lcs_set_allowed_threads(struct lcs_card *card, unsigned long threads)
312 spin_lock_irqsave(&card->mask_lock, flags);
313 card->thread_allowed_mask = threads;
314 spin_unlock_irqrestore(&card->mask_lock, flags);
315 wake_up(&card->wait_q);
317 static int lcs_threads_running(struct lcs_card *card, unsigned long threads)
322 spin_lock_irqsave(&card->mask_lock, flags);
323 rc = (card->thread_running_mask & threads);
324 spin_unlock_irqrestore(&card->mask_lock, flags);
329 lcs_wait_for_threads(struct lcs_card *card, unsigned long threads)
331 return wait_event_interruptible(card->wait_q,
332 lcs_threads_running(card, threads) == 0);
335 static int lcs_set_thread_start_bit(struct lcs_card *card, unsigned long thread)
339 spin_lock_irqsave(&card->mask_lock, flags);
340 if ( !(card->thread_allowed_mask & thread) ||
341 (card->thread_start_mask & thread) ) {
342 spin_unlock_irqrestore(&card->mask_lock, flags);
345 card->thread_start_mask |= thread;
346 spin_unlock_irqrestore(&card->mask_lock, flags);
351 lcs_clear_thread_running_bit(struct lcs_card *card, unsigned long thread)
355 spin_lock_irqsave(&card->mask_lock, flags);
356 card->thread_running_mask &= ~thread;
357 spin_unlock_irqrestore(&card->mask_lock, flags);
358 wake_up(&card->wait_q);
361 static int __lcs_do_run_thread(struct lcs_card *card, unsigned long thread)
366 spin_lock_irqsave(&card->mask_lock, flags);
367 if (card->thread_start_mask & thread){
368 if ((card->thread_allowed_mask & thread) &&
369 !(card->thread_running_mask & thread)){
371 card->thread_start_mask &= ~thread;
372 card->thread_running_mask |= thread;
376 spin_unlock_irqrestore(&card->mask_lock, flags);
381 lcs_do_run_thread(struct lcs_card *card, unsigned long thread)
384 wait_event(card->wait_q,
385 (rc = __lcs_do_run_thread(card, thread)) >= 0);
390 lcs_do_start_thread(struct lcs_card *card, unsigned long thread)
395 spin_lock_irqsave(&card->mask_lock, flags);
397 (u8) card->thread_start_mask,
398 (u8) card->thread_allowed_mask,
399 (u8) card->thread_running_mask);
400 rc = (card->thread_start_mask & thread);
401 spin_unlock_irqrestore(&card->mask_lock, flags);
406 * Initialize channels,card and state machines.
409 lcs_setup_card(struct lcs_card *card)
412 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
414 lcs_setup_read(card);
415 lcs_setup_write(card);
417 card->state = DEV_STATE_DOWN;
418 card->tx_buffer = NULL;
419 card->tx_emitted = 0;
421 init_waitqueue_head(&card->wait_q);
422 spin_lock_init(&card->lock);
423 spin_lock_init(&card->ipm_lock);
424 spin_lock_init(&card->mask_lock);
426 INIT_LIST_HEAD(&card->ipm_list);
428 INIT_LIST_HEAD(&card->lancmd_waiters);
431 static void lcs_clear_multicast_list(struct lcs_card *card)
439 spin_lock_irqsave(&card->ipm_lock, flags);
440 while (!list_empty(&card->ipm_list)){
441 ipm = list_entry(card->ipm_list.next,
445 spin_unlock_irqrestore(&card->ipm_lock, flags);
446 lcs_send_delipm(card, ipm);
447 spin_lock_irqsave(&card->ipm_lock, flags);
451 spin_unlock_irqrestore(&card->ipm_lock, flags);
455 * Cleanup channels,card and state machines.
458 lcs_cleanup_card(struct lcs_card *card)
462 LCS_DBF_HEX(2,setup,&card,sizeof(void*));
464 if (card->dev != NULL)
465 free_netdev(card->dev);
467 lcs_cleanup_channel(&card->write);
468 lcs_cleanup_channel(&card->read);
552 lcs_start_channels(struct lcs_card *card)
558 rc = lcs_start_channel(&card->read);
562 rc = lcs_start_channel(&card->write);
564 lcs_stop_channel(&card->read);
572 lcs_stop_channels(struct lcs_card *card)
575 lcs_stop_channel(&card->read);
576 lcs_stop_channel(&card->write);
735 lcs_get_lancmd(struct lcs_card *card, int count)
742 wait_event(card->write.wait_q,
743 ((buffer = lcs_get_buffer(&card->write)) != NULL));
792 lcs_notify_lancmd_waiters(struct lcs_card *card, struct lcs_cmd *cmd)
798 spin_lock(&card->lock);
799 list_for_each_safe(l, n, &card->lancmd_waiters) {
805 reply->callback(card, cmd);
813 spin_unlock(&card->lock);
827 spin_lock_irqsave(&reply->card->lock, flags);
829 &reply->card->lancmd_waiters,list) {
833 spin_unlock_irqrestore(&reply->card->lock, flags);
841 spin_unlock_irqrestore(&reply->card->lock, flags);
845 lcs_send_lancmd(struct lcs_card *card, struct lcs_buffer *buffer,
856 cmd->sequence_no = card->sequence_no++;
861 reply->card = card;
862 spin_lock_irqsave(&card->lock, flags);
863 list_add_tail(&reply->list, &card->lancmd_waiters);
864 spin_unlock_irqrestore(&card->lock, flags);
867 rc = lcs_ready_buffer(&card->write, buffer);
871 mod_timer(&reply->timer, jiffies + HZ * card->lancmd_timeout);
884 lcs_send_startup(struct lcs_card *card, __u8 initiator)
890 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
895 return lcs_send_lancmd(card, buffer, NULL);
902 lcs_send_shutdown(struct lcs_card *card)
908 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
912 return lcs_send_lancmd(card, buffer, NULL);
919 __lcs_lanstat_cb(struct lcs_card *card, struct lcs_cmd *cmd)
922 memcpy(card->mac, cmd->cmd.lcs_lanstat_cmd.mac_addr, LCS_MAC_LENGTH);
926 lcs_send_lanstat(struct lcs_card *card)
932 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
937 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
938 cmd->cmd.lcs_std_cmd.portno = card->portno;
939 return lcs_send_lancmd(card, buffer, __lcs_lanstat_cb);
946 lcs_send_stoplan(struct lcs_card *card, __u8 initiator)
952 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
956 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
957 cmd->cmd.lcs_std_cmd.portno = card->portno;
958 return lcs_send_lancmd(card, buffer, NULL);
965 __lcs_send_startlan_cb(struct lcs_card *card, struct lcs_cmd *cmd)
968 card->lan_type = cmd->cmd.lcs_std_cmd.lan_type;
969 card->portno = cmd->cmd.lcs_std_cmd.portno;
973 lcs_send_startlan(struct lcs_card *card, __u8 initiator)
979 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
983 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
984 cmd->cmd.lcs_std_cmd.portno = card->portno;
985 return lcs_send_lancmd(card, buffer, __lcs_send_startlan_cb);
993 lcs_send_setipm(struct lcs_card *card,struct lcs_ipm_list *ipm_list)
999 buffer = lcs_get_lancmd(card, LCS_MULTICAST_CMD_SIZE);
1003 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
1004 cmd->cmd.lcs_qipassist.portno = card->portno;
1010 return lcs_send_lancmd(card, buffer, NULL);
1017 lcs_send_delipm(struct lcs_card *card,struct lcs_ipm_list *ipm_list)
1023 buffer = lcs_get_lancmd(card, LCS_MULTICAST_CMD_SIZE);
1027 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
1028 cmd->cmd.lcs_qipassist.portno = card->portno;
1034 return lcs_send_lancmd(card, buffer, NULL);
1041 __lcs_check_multicast_cb(struct lcs_card *card, struct lcs_cmd *cmd)
1044 card->ip_assists_supported =
1046 card->ip_assists_enabled =
1051 lcs_check_multicast_support(struct lcs_card *card)
1059 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
1063 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
1064 cmd->cmd.lcs_qipassist.portno = card->portno;
1067 rc = lcs_send_lancmd(card, buffer, __lcs_check_multicast_cb);
1072 if (card->ip_assists_supported & LCS_IPASS_MULTICAST_SUPPORT)
1078 * set or del multicast address on LCS card
1081 lcs_fix_multicast_list(struct lcs_card *card)
1090 spin_lock_irqsave(&card->ipm_lock, flags);
1092 list_for_each_entry_safe(ipm, tmp, &card->ipm_list, list){
1098 spin_unlock_irqrestore(&card->ipm_lock, flags);
1099 rc = lcs_send_setipm(card, ipm);
1100 spin_lock_irqsave(&card->ipm_lock, flags);
1111 list_add_tail(&ipm->list, &card->ipm_list);
1116 spin_unlock_irqrestore(&card->ipm_lock, flags);
1117 lcs_send_delipm(card, ipm);
1118 spin_lock_irqsave(&card->ipm_lock, flags);
1127 list_move_tail(&ipm->list, &card->ipm_list);
1129 spin_unlock_irqrestore(&card->ipm_lock, flags);
1145 static void lcs_remove_mc_addresses(struct lcs_card *card,
1155 spin_lock_irqsave(&card->ipm_lock, flags);
1156 list_for_each(l, &card->ipm_list) {
1160 lcs_get_mac_for_ipm(im4->multiaddr, buf, card->dev);
1169 spin_unlock_irqrestore(&card->ipm_lock, flags);
1172 static struct lcs_ipm_list *lcs_check_addr_entry(struct lcs_card *card,
1181 spin_lock_irqsave(&card->ipm_lock, flags);
1182 list_for_each(l, &card->ipm_list) {
1191 spin_unlock_irqrestore(&card->ipm_lock, flags);
1195 static void lcs_set_mc_addresses(struct lcs_card *card,
1207 lcs_get_mac_for_ipm(im4->multiaddr, buf, card->dev);
1208 ipm = lcs_check_addr_entry(card, im4, buf);
1220 spin_lock_irqsave(&card->ipm_lock, flags);
1222 list_add(&ipm->list, &card->ipm_list);
1223 spin_unlock_irqrestore(&card->ipm_lock, flags);
1230 struct lcs_card *card;
1233 card = (struct lcs_card *) data;
1235 if (!lcs_do_run_thread(card, LCS_SET_MC_THREAD))
1239 in4_dev = in_dev_get(card->dev);
1243 lcs_remove_mc_addresses(card,in4_dev);
1244 lcs_set_mc_addresses(card, in4_dev);
1248 netif_carrier_off(card->dev);
1249 netif_tx_disable(card->dev);
1250 wait_event(card->write.wait_q,
1251 (card->write.state != LCS_CH_STATE_RUNNING));
1252 lcs_fix_multicast_list(card);
1253 if (card->state == DEV_STATE_UP) {
1254 netif_carrier_on(card->dev);
1255 netif_wake_queue(card->dev);
1258 lcs_clear_thread_running_bit(card, LCS_SET_MC_THREAD);
1271 struct lcs_card *card;
1274 card = (struct lcs_card *) dev->ml_priv;
1276 if (!lcs_set_thread_start_bit(card, LCS_SET_MC_THREAD))
1277 schedule_work(&card->kernel_thread_starter);
1351 lcs_schedule_recovery(struct lcs_card *card)
1354 if (!lcs_set_thread_start_bit(card, LCS_RECOVERY_THREAD))
1355 schedule_work(&card->kernel_thread_starter);
1364 struct lcs_card *card;
1372 card = CARD_FROM_DEV(cdev);
1373 if (card->read.ccwdev == cdev)
1374 channel = &card->read;
1376 channel = &card->write;
1398 lcs_schedule_recovery(card);
1399 wake_up(&card->wait_q);
1483 __lcs_emit_txbuffer(struct lcs_card *card)
1486 *(__u16 *)(card->tx_buffer->data + card->tx_buffer->count) = 0;
1487 card->tx_buffer->count += 2;
1488 lcs_ready_buffer(&card->write, card->tx_buffer);
1489 card->tx_buffer = NULL;
1490 card->tx_emitted++;
1499 struct lcs_card *card;
1504 card = container_of(channel, struct lcs_card, write);
1505 if (netif_queue_stopped(card->dev) && netif_carrier_ok(card->dev))
1506 netif_wake_queue(card->dev);
1507 spin_lock(&card->lock);
1508 card->tx_emitted--;
1509 if (card->tx_emitted <= 0 && card->tx_buffer != NULL)
1514 __lcs_emit_txbuffer(card);
1515 spin_unlock(&card->lock);
1521 static netdev_tx_t __lcs_start_xmit(struct lcs_card *card, struct sk_buff *skb,
1529 card->stats.tx_dropped++;
1530 card->stats.tx_errors++;
1533 if (card->state != DEV_STATE_UP) {
1535 card->stats.tx_dropped++;
1536 card->stats.tx_errors++;
1537 card->stats.tx_carrier_errors++;
1544 netif_stop_queue(card->dev);
1545 spin_lock(&card->lock);
1546 if (card->tx_buffer != NULL &&
1547 card->tx_buffer->count + sizeof(struct lcs_header) +
1550 __lcs_emit_txbuffer(card);
1551 if (card->tx_buffer == NULL) {
1553 card->tx_buffer = lcs_get_buffer(&card->write);
1554 if (card->tx_buffer == NULL) {
1555 card->stats.tx_dropped++;
1559 card->tx_buffer->callback = lcs_txbuffer_cb;
1560 card->tx_buffer->count = 0;
1563 (card->tx_buffer->data + card->tx_buffer->count);
1564 card->tx_buffer->count += skb->len + sizeof(struct lcs_header);
1565 header->offset = card->tx_buffer->count;
1566 header->type = card->lan_type;
1567 header->slot = card->portno;
1569 spin_unlock(&card->lock);
1570 card->stats.tx_bytes += skb->len;
1571 card->stats.tx_packets++;
1573 netif_wake_queue(card->dev);
1574 spin_lock(&card->lock);
1575 if (card->tx_emitted <= 0 && card->tx_buffer != NULL)
1577 __lcs_emit_txbuffer(card);
1579 spin_unlock(&card->lock);
1585 struct lcs_card *card;
1589 card = (struct lcs_card *) dev->ml_priv;
1590 rc = __lcs_start_xmit(card, skb, dev);
1598 lcs_startlan_auto(struct lcs_card *card)
1604 card->lan_type = LCS_FRAME_TYPE_ENET;
1605 rc = lcs_send_startlan(card, LCS_INITIATOR_TCPIP);
1611 card->lan_type = LCS_FRAME_TYPE_FDDI;
1612 rc = lcs_send_startlan(card, LCS_INITIATOR_TCPIP);
1620 lcs_startlan(struct lcs_card *card)
1626 if (card->portno != LCS_INVALID_PORT_NO) {
1627 if (card->lan_type == LCS_FRAME_TYPE_AUTO)
1628 rc = lcs_startlan_auto(card);
1630 rc = lcs_send_startlan(card, LCS_INITIATOR_TCPIP);
1633 card->portno = i;
1634 if (card->lan_type != LCS_FRAME_TYPE_AUTO)
1635 rc = lcs_send_startlan(card,
1639 rc = lcs_startlan_auto(card);
1645 return lcs_send_lanstat(card);
1654 lcs_detect(struct lcs_card *card)
1659 /* start/reset card */
1660 if (card->dev)
1661 netif_stop_queue(card->dev);
1662 rc = lcs_stop_channels(card);
1664 rc = lcs_start_channels(card);
1666 rc = lcs_send_startup(card, LCS_INITIATOR_TCPIP);
1668 rc = lcs_startlan(card);
1672 card->state = DEV_STATE_UP;
1674 card->state = DEV_STATE_DOWN;
1675 card->write.state = LCS_CH_STATE_INIT;
1676 card->read.state = LCS_CH_STATE_INIT;
1682 * LCS Stop card
1685 lcs_stopcard(struct lcs_card *card)
1691 if (card->read.state != LCS_CH_STATE_STOPPED &&
1692 card->write.state != LCS_CH_STATE_STOPPED &&
1693 card->read.state != LCS_CH_STATE_ERROR &&
1694 card->write.state != LCS_CH_STATE_ERROR &&
1695 card->state == DEV_STATE_UP) {
1696 lcs_clear_multicast_list(card);
1697 rc = lcs_send_stoplan(card,LCS_INITIATOR_TCPIP);
1698 rc = lcs_send_shutdown(card);
1700 rc = lcs_stop_channels(card);
1701 card->state = DEV_STATE_DOWN;
1712 struct lcs_card *card = container_of(work, struct lcs_card, kernel_thread_starter);
1714 if (lcs_do_start_thread(card, LCS_RECOVERY_THREAD))
1715 kthread_run(lcs_recovery, card, "lcs_recover");
1717 if (lcs_do_start_thread(card, LCS_SET_MC_THREAD))
1718 kthread_run(lcs_register_mc_addresses, card, "regipm");
1726 lcs_get_control(struct lcs_card *card, struct lcs_cmd *cmd)
1733 lcs_schedule_recovery(card);
1736 if (card->dev) {
1738 card->dev->name);
1739 netif_carrier_off(card->dev);
1747 lcs_notify_lancmd_waiters(card, cmd);
1754 lcs_get_skb(struct lcs_card *card, char *skb_data, unsigned int skb_len)
1759 if (card->dev == NULL ||
1760 card->state != DEV_STATE_UP)
1761 /* The card isn't up. Ignore the packet. */
1766 dev_err(&card->dev->dev,
1768 card->dev->name);
1769 card->stats.rx_dropped++;
1773 skb->protocol = card->lan_type_trans(skb, card->dev);
1774 card->stats.rx_bytes += skb_len;
1775 card->stats.rx_packets++;
1777 *((__u32 *)skb->cb) = ++card->pkt_seq;
1787 struct lcs_card *card;
1797 card = container_of(channel, struct lcs_card, read);
1804 card->stats.rx_length_errors++;
1805 card->stats.rx_errors++;
1811 lcs_get_control(card, (struct lcs_cmd *) lcs_hdr);
1816 lcs_get_skb(card, (char *)(lcs_hdr + 1),
1828 lcs_ready_buffer(&card->read, buffer);
1837 struct lcs_card *card;
1840 card = (struct lcs_card *) dev->ml_priv;
1841 return &card->stats;
1851 struct lcs_card *card;
1855 card = (struct lcs_card *) dev->ml_priv;
1859 wait_event(card->write.wait_q,
1860 (card->write.state != LCS_CH_STATE_RUNNING));
1861 rc = lcs_stopcard(card);
1863 dev_err(&card->dev->dev,
1875 struct lcs_card *card;
1879 card = (struct lcs_card *) dev->ml_priv;
1881 rc = lcs_detect(card);
1889 card->state = DEV_STATE_UP;
1900 struct lcs_card *card;
1902 card = dev_get_drvdata(dev);
1904 if (!card)
1907 return sprintf(buf, "%d\n", card->portno);
1916 struct lcs_card *card;
1920 card = dev_get_drvdata(dev);
1922 if (!card)
1929 card->portno = value;
1930 if (card->dev)
1931 card->dev->dev_port = card->portno;
1943 "OSA LCS card",
1965 struct lcs_card *card;
1967 card = dev_get_drvdata(dev);
1969 return card ? sprintf(buf, "%u\n", card->lancmd_timeout) : 0;
1975 struct lcs_card *card;
1979 card = dev_get_drvdata(dev);
1981 if (!card)
1988 card->lancmd_timeout = value;
2000 struct lcs_card *card = dev_get_drvdata(dev);
2004 if (!card)
2006 if (card->state != DEV_STATE_UP)
2010 lcs_schedule_recovery(card);
2041 struct lcs_card *card;
2047 card = lcs_alloc_card();
2048 if (!card) {
2053 dev_set_drvdata(&ccwgdev->dev, card);
2056 card->gdev = ccwgdev;
2057 INIT_WORK(&card->kernel_thread_starter, lcs_start_kernel_thread);
2058 card->thread_start_mask = 0;
2059 card->thread_allowed_mask = 0;
2060 card->thread_running_mask = 0;
2069 struct lcs_card *card;
2072 card = dev_get_drvdata(&ccwgdev->dev);
2073 if (card->dev->reg_state != NETREG_UNINITIALIZED)
2075 SET_NETDEV_DEV(card->dev, &ccwgdev->dev);
2076 return register_netdev(card->dev);
2100 struct lcs_card *card;
2105 card = dev_get_drvdata(&ccwgdev->dev);
2106 if (!card)
2110 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2111 card->read.ccwdev = ccwgdev->cdev[0];
2112 card->write.ccwdev = ccwgdev->cdev[1];
2114 recover_state = card->state;
2115 rc = ccw_device_set_online(card->read.ccwdev);
2118 rc = ccw_device_set_online(card->write.ccwdev);
2124 lcs_setup_card(card);
2125 rc = lcs_detect(card);
2131 lcs_stopcard(card);
2134 if (card->dev) {
2136 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2139 switch (card->lan_type) {
2142 card->lan_type_trans = eth_type_trans;
2148 card->lan_type_trans = fddi_type_trans;
2159 card->dev = dev;
2160 card->dev->ml_priv = card;
2161 card->dev->netdev_ops = &lcs_netdev_ops;
2162 card->dev->dev_port = card->portno;
2163 memcpy(card->dev->dev_addr, card->mac, LCS_MAC_LENGTH);
2165 if (!lcs_check_multicast_support(card))
2166 card->dev->netdev_ops = &lcs_mc_netdev_ops;
2169 lcs_set_allowed_threads(card,0xffffffff);
2171 lcs_set_multicast_list(card->dev);
2172 card->dev->flags |= IFF_UP;
2173 netif_carrier_on(card->dev);
2174 netif_wake_queue(card->dev);
2175 card->state = DEV_STATE_UP;
2177 lcs_stopcard(card);
2184 pr_info("LCS device %s %s IPv6 support\n", card->dev->name,
2185 (card->ip_assists_supported & LCS_IPASS_IPV6_SUPPORT) ?
2188 pr_info("LCS device %s %s Multicast support\n", card->dev->name,
2189 (card->ip_assists_supported & LCS_IPASS_MULTICAST_SUPPORT) ?
2194 ccw_device_set_offline(card->write.ccwdev);
2196 ccw_device_set_offline(card->read.ccwdev);
2207 struct lcs_card *card;
2212 card = dev_get_drvdata(&ccwgdev->dev);
2213 if (!card)
2216 lcs_set_allowed_threads(card, 0);
2217 if (lcs_wait_for_threads(card, LCS_SET_MC_THREAD))
2220 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2221 recover_state = card->state;
2223 ret = lcs_stop_device(card->dev);
2224 ret2 = ccw_device_set_offline(card->read.ccwdev);
2225 ret3 = ccw_device_set_offline(card->write.ccwdev);
2231 card->state = DEV_STATE_RECOVER;
2248 struct lcs_card *card;
2252 card = (struct lcs_card *) ptr;
2255 if (!lcs_do_run_thread(card, LCS_RECOVERY_THREAD))
2258 gdev = card->gdev;
2265 card->dev->name);
2268 card->dev->name);
2269 lcs_clear_thread_running_bit(card, LCS_RECOVERY_THREAD);
2274 * lcs_remove_device, free buffers and card
2279 struct lcs_card *card;
2281 card = dev_get_drvdata(&ccwgdev->dev);
2282 if (!card)
2286 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2290 if (card->dev)
2291 unregister_netdev(card->dev);
2292 lcs_cleanup_card(card);
2293 lcs_free_card(card);