Lines Matching defs:card

150  * LCS free memory for card and channels.
153 lcs_free_card(struct lcs_card *card)
156 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
157 kfree(card);
161 * LCS alloc memory for card and channels
166 struct lcs_card *card;
171 card = kzalloc(sizeof(struct lcs_card), GFP_KERNEL | GFP_DMA);
172 if (card == NULL)
174 card->lan_type = LCS_FRAME_TYPE_AUTO;
175 card->pkt_seq = 0;
176 card->lancmd_timeout = LCS_LANCMD_TIMEOUT_DEFAULT;
178 rc = lcs_alloc_channel(&card->read);
181 lcs_free_card(card);
185 rc = lcs_alloc_channel(&card->write);
188 lcs_cleanup_channel(&card->read);
189 lcs_free_card(card);
194 INIT_LIST_HEAD(&card->ipm_list);
196 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
197 return card;
204 lcs_setup_read_ccws(struct lcs_card *card)
210 memset(card->read.ccws, 0, sizeof (struct ccw1) * (LCS_NUM_BUFFS + 1));
212 card->read.ccws[cnt].cmd_code = LCS_CCW_READ;
213 card->read.ccws[cnt].count = LCS_IOBUFFERSIZE;
214 card->read.ccws[cnt].flags =
220 card->read.ccws[cnt].cda =
221 virt_to_dma32(card->read.iob[cnt].data);
223 card->read.iob[cnt].data)->offset = LCS_ILLEGAL_OFFSET;
224 card->read.iob[cnt].callback = lcs_get_frames_cb;
225 card->read.iob[cnt].state = LCS_BUF_STATE_READY;
226 card->read.iob[cnt].count = LCS_IOBUFFERSIZE;
228 card->read.ccws[0].flags &= ~CCW_FLAG_PCI;
229 card->read.ccws[LCS_NUM_BUFFS - 1].flags &= ~CCW_FLAG_PCI;
230 card->read.ccws[LCS_NUM_BUFFS - 1].flags |= CCW_FLAG_SUSPEND;
232 card->read.ccws[LCS_NUM_BUFFS].cmd_code = LCS_CCW_TRANSFER;
233 card->read.ccws[LCS_NUM_BUFFS].cda = virt_to_dma32(card->read.ccws);
235 card->read.state = LCS_CH_STATE_INIT;
237 card->read.io_idx = 0;
238 card->read.buf_idx = 0;
242 lcs_setup_read(struct lcs_card *card)
246 lcs_setup_read_ccws(card);
248 card->read.irq_tasklet.data = (unsigned long) &card->read;
249 card->read.irq_tasklet.func = lcs_tasklet;
251 init_waitqueue_head(&card->read.wait_q);
258 lcs_setup_write_ccws(struct lcs_card *card)
264 memset(card->write.ccws, 0, sizeof(struct ccw1) * (LCS_NUM_BUFFS + 1));
266 card->write.ccws[cnt].cmd_code = LCS_CCW_WRITE;
267 card->write.ccws[cnt].count = 0;
268 card->write.ccws[cnt].flags =
274 card->write.ccws[cnt].cda =
275 virt_to_dma32(card->write.iob[cnt].data);
278 card->write.ccws[LCS_NUM_BUFFS].cmd_code = LCS_CCW_TRANSFER;
279 card->write.ccws[LCS_NUM_BUFFS].cda = virt_to_dma32(card->write.ccws);
281 card->read.state = LCS_CH_STATE_INIT;
283 card->write.io_idx = 0;
284 card->write.buf_idx = 0;
288 lcs_setup_write(struct lcs_card *card)
292 lcs_setup_write_ccws(card);
294 card->write.irq_tasklet.data = (unsigned long) &card->write;
295 card->write.irq_tasklet.func = lcs_tasklet;
297 init_waitqueue_head(&card->write.wait_q);
301 lcs_set_allowed_threads(struct lcs_card *card, unsigned long threads)
305 spin_lock_irqsave(&card->mask_lock, flags);
306 card->thread_allowed_mask = threads;
307 spin_unlock_irqrestore(&card->mask_lock, flags);
308 wake_up(&card->wait_q);
310 static int lcs_threads_running(struct lcs_card *card, unsigned long threads)
315 spin_lock_irqsave(&card->mask_lock, flags);
316 rc = (card->thread_running_mask & threads);
317 spin_unlock_irqrestore(&card->mask_lock, flags);
322 lcs_wait_for_threads(struct lcs_card *card, unsigned long threads)
324 return wait_event_interruptible(card->wait_q,
325 lcs_threads_running(card, threads) == 0);
328 static int lcs_set_thread_start_bit(struct lcs_card *card, unsigned long thread)
332 spin_lock_irqsave(&card->mask_lock, flags);
333 if ( !(card->thread_allowed_mask & thread) ||
334 (card->thread_start_mask & thread) ) {
335 spin_unlock_irqrestore(&card->mask_lock, flags);
338 card->thread_start_mask |= thread;
339 spin_unlock_irqrestore(&card->mask_lock, flags);
344 lcs_clear_thread_running_bit(struct lcs_card *card, unsigned long thread)
348 spin_lock_irqsave(&card->mask_lock, flags);
349 card->thread_running_mask &= ~thread;
350 spin_unlock_irqrestore(&card->mask_lock, flags);
351 wake_up(&card->wait_q);
354 static int __lcs_do_run_thread(struct lcs_card *card, unsigned long thread)
359 spin_lock_irqsave(&card->mask_lock, flags);
360 if (card->thread_start_mask & thread){
361 if ((card->thread_allowed_mask & thread) &&
362 !(card->thread_running_mask & thread)){
364 card->thread_start_mask &= ~thread;
365 card->thread_running_mask |= thread;
369 spin_unlock_irqrestore(&card->mask_lock, flags);
374 lcs_do_run_thread(struct lcs_card *card, unsigned long thread)
377 wait_event(card->wait_q,
378 (rc = __lcs_do_run_thread(card, thread)) >= 0);
383 lcs_do_start_thread(struct lcs_card *card, unsigned long thread)
388 spin_lock_irqsave(&card->mask_lock, flags);
390 (u8) card->thread_start_mask,
391 (u8) card->thread_allowed_mask,
392 (u8) card->thread_running_mask);
393 rc = (card->thread_start_mask & thread);
394 spin_unlock_irqrestore(&card->mask_lock, flags);
399 * Initialize channels,card and state machines.
402 lcs_setup_card(struct lcs_card *card)
405 LCS_DBF_HEX(2, setup, &card, sizeof(void*));
407 lcs_setup_read(card);
408 lcs_setup_write(card);
410 card->state = DEV_STATE_DOWN;
411 card->tx_buffer = NULL;
412 card->tx_emitted = 0;
414 init_waitqueue_head(&card->wait_q);
415 spin_lock_init(&card->lock);
416 spin_lock_init(&card->ipm_lock);
417 spin_lock_init(&card->mask_lock);
419 INIT_LIST_HEAD(&card->ipm_list);
421 INIT_LIST_HEAD(&card->lancmd_waiters);
424 static void lcs_clear_multicast_list(struct lcs_card *card)
432 spin_lock_irqsave(&card->ipm_lock, flags);
433 while (!list_empty(&card->ipm_list)){
434 ipm = list_entry(card->ipm_list.next,
438 spin_unlock_irqrestore(&card->ipm_lock, flags);
439 lcs_send_delipm(card, ipm);
440 spin_lock_irqsave(&card->ipm_lock, flags);
444 spin_unlock_irqrestore(&card->ipm_lock, flags);
449 * Cleanup channels,card and state machines.
452 lcs_cleanup_card(struct lcs_card *card)
456 LCS_DBF_HEX(2,setup,&card,sizeof(void*));
458 if (card->dev != NULL)
459 free_netdev(card->dev);
461 lcs_cleanup_channel(&card->write);
462 lcs_cleanup_channel(&card->read);
546 lcs_start_channels(struct lcs_card *card)
552 rc = lcs_start_channel(&card->read);
556 rc = lcs_start_channel(&card->write);
558 lcs_stop_channel(&card->read);
566 lcs_stop_channels(struct lcs_card *card)
569 lcs_stop_channel(&card->read);
570 lcs_stop_channel(&card->write);
729 lcs_get_lancmd(struct lcs_card *card, int count)
736 wait_event(card->write.wait_q,
737 ((buffer = lcs_get_buffer(&card->write)) != NULL));
786 lcs_notify_lancmd_waiters(struct lcs_card *card, struct lcs_cmd *cmd)
792 spin_lock(&card->lock);
793 list_for_each_safe(l, n, &card->lancmd_waiters) {
799 reply->callback(card, cmd);
807 spin_unlock(&card->lock);
821 spin_lock_irqsave(&reply->card->lock, flags);
823 &reply->card->lancmd_waiters,list) {
827 spin_unlock_irqrestore(&reply->card->lock, flags);
835 spin_unlock_irqrestore(&reply->card->lock, flags);
839 lcs_send_lancmd(struct lcs_card *card, struct lcs_buffer *buffer,
850 cmd->sequence_no = card->sequence_no++;
855 reply->card = card;
856 spin_lock_irqsave(&card->lock, flags);
857 list_add_tail(&reply->list, &card->lancmd_waiters);
858 spin_unlock_irqrestore(&card->lock, flags);
861 rc = lcs_ready_buffer(&card->write, buffer);
865 mod_timer(&reply->timer, jiffies + HZ * card->lancmd_timeout);
878 lcs_send_startup(struct lcs_card *card, __u8 initiator)
884 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
889 return lcs_send_lancmd(card, buffer, NULL);
896 lcs_send_shutdown(struct lcs_card *card)
902 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
906 return lcs_send_lancmd(card, buffer, NULL);
913 __lcs_lanstat_cb(struct lcs_card *card, struct lcs_cmd *cmd)
916 memcpy(card->mac, cmd->cmd.lcs_lanstat_cmd.mac_addr, LCS_MAC_LENGTH);
920 lcs_send_lanstat(struct lcs_card *card)
926 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
931 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
932 cmd->cmd.lcs_std_cmd.portno = card->portno;
933 return lcs_send_lancmd(card, buffer, __lcs_lanstat_cb);
940 lcs_send_stoplan(struct lcs_card *card, __u8 initiator)
946 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
950 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
951 cmd->cmd.lcs_std_cmd.portno = card->portno;
952 return lcs_send_lancmd(card, buffer, NULL);
959 __lcs_send_startlan_cb(struct lcs_card *card, struct lcs_cmd *cmd)
962 card->lan_type = cmd->cmd.lcs_std_cmd.lan_type;
963 card->portno = cmd->cmd.lcs_std_cmd.portno;
967 lcs_send_startlan(struct lcs_card *card, __u8 initiator)
973 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
977 cmd->cmd.lcs_std_cmd.lan_type = card->lan_type;
978 cmd->cmd.lcs_std_cmd.portno = card->portno;
979 return lcs_send_lancmd(card, buffer, __lcs_send_startlan_cb);
987 lcs_send_setipm(struct lcs_card *card,struct lcs_ipm_list *ipm_list)
993 buffer = lcs_get_lancmd(card, LCS_MULTICAST_CMD_SIZE);
997 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
998 cmd->cmd.lcs_qipassist.portno = card->portno;
1004 return lcs_send_lancmd(card, buffer, NULL);
1011 lcs_send_delipm(struct lcs_card *card,struct lcs_ipm_list *ipm_list)
1017 buffer = lcs_get_lancmd(card, LCS_MULTICAST_CMD_SIZE);
1021 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
1022 cmd->cmd.lcs_qipassist.portno = card->portno;
1028 return lcs_send_lancmd(card, buffer, NULL);
1035 __lcs_check_multicast_cb(struct lcs_card *card, struct lcs_cmd *cmd)
1038 card->ip_assists_supported =
1040 card->ip_assists_enabled =
1045 lcs_check_multicast_support(struct lcs_card *card)
1053 buffer = lcs_get_lancmd(card, LCS_STD_CMD_SIZE);
1057 cmd->cmd.lcs_qipassist.lan_type = card->lan_type;
1058 cmd->cmd.lcs_qipassist.portno = card->portno;
1061 rc = lcs_send_lancmd(card, buffer, __lcs_check_multicast_cb);
1066 if (card->ip_assists_supported & LCS_IPASS_MULTICAST_SUPPORT)
1072 * set or del multicast address on LCS card
1075 lcs_fix_multicast_list(struct lcs_card *card)
1084 spin_lock_irqsave(&card->ipm_lock, flags);
1086 list_for_each_entry_safe(ipm, tmp, &card->ipm_list, list){
1092 spin_unlock_irqrestore(&card->ipm_lock, flags);
1093 rc = lcs_send_setipm(card, ipm);
1094 spin_lock_irqsave(&card->ipm_lock, flags);
1105 list_add_tail(&ipm->list, &card->ipm_list);
1110 spin_unlock_irqrestore(&card->ipm_lock, flags);
1111 lcs_send_delipm(card, ipm);
1112 spin_lock_irqsave(&card->ipm_lock, flags);
1121 list_move_tail(&ipm->list, &card->ipm_list);
1123 spin_unlock_irqrestore(&card->ipm_lock, flags);
1139 static void lcs_remove_mc_addresses(struct lcs_card *card,
1149 spin_lock_irqsave(&card->ipm_lock, flags);
1150 list_for_each(l, &card->ipm_list) {
1154 lcs_get_mac_for_ipm(im4->multiaddr, buf, card->dev);
1163 spin_unlock_irqrestore(&card->ipm_lock, flags);
1166 static struct lcs_ipm_list *lcs_check_addr_entry(struct lcs_card *card,
1175 spin_lock_irqsave(&card->ipm_lock, flags);
1176 list_for_each(l, &card->ipm_list) {
1185 spin_unlock_irqrestore(&card->ipm_lock, flags);
1189 static void lcs_set_mc_addresses(struct lcs_card *card,
1201 lcs_get_mac_for_ipm(im4->multiaddr, buf, card->dev);
1202 ipm = lcs_check_addr_entry(card, im4, buf);
1214 spin_lock_irqsave(&card->ipm_lock, flags);
1216 list_add(&ipm->list, &card->ipm_list);
1217 spin_unlock_irqrestore(&card->ipm_lock, flags);
1224 struct lcs_card *card;
1227 card = (struct lcs_card *) data;
1229 if (!lcs_do_run_thread(card, LCS_SET_MC_THREAD))
1233 in4_dev = in_dev_get(card->dev);
1237 lcs_remove_mc_addresses(card,in4_dev);
1238 lcs_set_mc_addresses(card, in4_dev);
1242 netif_carrier_off(card->dev);
1243 netif_tx_disable(card->dev);
1244 wait_event(card->write.wait_q,
1245 (card->write.state != LCS_CH_STATE_RUNNING));
1246 lcs_fix_multicast_list(card);
1247 if (card->state == DEV_STATE_UP) {
1248 netif_carrier_on(card->dev);
1249 netif_wake_queue(card->dev);
1252 lcs_clear_thread_running_bit(card, LCS_SET_MC_THREAD);
1265 struct lcs_card *card;
1268 card = (struct lcs_card *) dev->ml_priv;
1270 if (!lcs_set_thread_start_bit(card, LCS_SET_MC_THREAD))
1271 schedule_work(&card->kernel_thread_starter);
1345 lcs_schedule_recovery(struct lcs_card *card)
1348 if (!lcs_set_thread_start_bit(card, LCS_RECOVERY_THREAD))
1349 schedule_work(&card->kernel_thread_starter);
1358 struct lcs_card *card;
1366 card = CARD_FROM_DEV(cdev);
1367 if (card->read.ccwdev == cdev)
1368 channel = &card->read;
1370 channel = &card->write;
1392 lcs_schedule_recovery(card);
1393 wake_up(&card->wait_q);
1477 __lcs_emit_txbuffer(struct lcs_card *card)
1480 *(__u16 *)(card->tx_buffer->data + card->tx_buffer->count) = 0;
1481 card->tx_buffer->count += 2;
1482 lcs_ready_buffer(&card->write, card->tx_buffer);
1483 card->tx_buffer = NULL;
1484 card->tx_emitted++;
1493 struct lcs_card *card;
1498 card = container_of(channel, struct lcs_card, write);
1499 if (netif_queue_stopped(card->dev) && netif_carrier_ok(card->dev))
1500 netif_wake_queue(card->dev);
1501 spin_lock(&card->lock);
1502 card->tx_emitted--;
1503 if (card->tx_emitted <= 0 && card->tx_buffer != NULL)
1508 __lcs_emit_txbuffer(card);
1509 spin_unlock(&card->lock);
1515 static netdev_tx_t __lcs_start_xmit(struct lcs_card *card, struct sk_buff *skb,
1523 card->stats.tx_dropped++;
1524 card->stats.tx_errors++;
1527 if (card->state != DEV_STATE_UP) {
1529 card->stats.tx_dropped++;
1530 card->stats.tx_errors++;
1531 card->stats.tx_carrier_errors++;
1538 netif_stop_queue(card->dev);
1539 spin_lock(&card->lock);
1540 if (card->tx_buffer != NULL &&
1541 card->tx_buffer->count + sizeof(struct lcs_header) +
1544 __lcs_emit_txbuffer(card);
1545 if (card->tx_buffer == NULL) {
1547 card->tx_buffer = lcs_get_buffer(&card->write);
1548 if (card->tx_buffer == NULL) {
1549 card->stats.tx_dropped++;
1553 card->tx_buffer->callback = lcs_txbuffer_cb;
1554 card->tx_buffer->count = 0;
1557 (card->tx_buffer->data + card->tx_buffer->count);
1558 card->tx_buffer->count += skb->len + sizeof(struct lcs_header);
1559 header->offset = card->tx_buffer->count;
1560 header->type = card->lan_type;
1561 header->slot = card->portno;
1563 spin_unlock(&card->lock);
1564 card->stats.tx_bytes += skb->len;
1565 card->stats.tx_packets++;
1567 netif_wake_queue(card->dev);
1568 spin_lock(&card->lock);
1569 if (card->tx_emitted <= 0 && card->tx_buffer != NULL)
1571 __lcs_emit_txbuffer(card);
1573 spin_unlock(&card->lock);
1579 struct lcs_card *card;
1583 card = (struct lcs_card *) dev->ml_priv;
1584 rc = __lcs_start_xmit(card, skb, dev);
1592 lcs_startlan_auto(struct lcs_card *card)
1597 card->lan_type = LCS_FRAME_TYPE_ENET;
1598 rc = lcs_send_startlan(card, LCS_INITIATOR_TCPIP);
1606 lcs_startlan(struct lcs_card *card)
1612 if (card->portno != LCS_INVALID_PORT_NO) {
1613 if (card->lan_type == LCS_FRAME_TYPE_AUTO)
1614 rc = lcs_startlan_auto(card);
1616 rc = lcs_send_startlan(card, LCS_INITIATOR_TCPIP);
1619 card->portno = i;
1620 if (card->lan_type != LCS_FRAME_TYPE_AUTO)
1621 rc = lcs_send_startlan(card,
1625 rc = lcs_startlan_auto(card);
1631 return lcs_send_lanstat(card);
1640 lcs_detect(struct lcs_card *card)
1645 /* start/reset card */
1646 if (card->dev)
1647 netif_stop_queue(card->dev);
1648 rc = lcs_stop_channels(card);
1650 rc = lcs_start_channels(card);
1652 rc = lcs_send_startup(card, LCS_INITIATOR_TCPIP);
1654 rc = lcs_startlan(card);
1658 card->state = DEV_STATE_UP;
1660 card->state = DEV_STATE_DOWN;
1661 card->write.state = LCS_CH_STATE_INIT;
1662 card->read.state = LCS_CH_STATE_INIT;
1668 * LCS Stop card
1671 lcs_stopcard(struct lcs_card *card)
1677 if (card->read.state != LCS_CH_STATE_STOPPED &&
1678 card->write.state != LCS_CH_STATE_STOPPED &&
1679 card->read.state != LCS_CH_STATE_ERROR &&
1680 card->write.state != LCS_CH_STATE_ERROR &&
1681 card->state == DEV_STATE_UP) {
1682 lcs_clear_multicast_list(card);
1683 rc = lcs_send_stoplan(card,LCS_INITIATOR_TCPIP);
1684 rc = lcs_send_shutdown(card);
1686 rc = lcs_stop_channels(card);
1687 card->state = DEV_STATE_DOWN;
1698 struct lcs_card *card = container_of(work, struct lcs_card, kernel_thread_starter);
1700 if (lcs_do_start_thread(card, LCS_RECOVERY_THREAD))
1701 kthread_run(lcs_recovery, card, "lcs_recover");
1703 if (lcs_do_start_thread(card, LCS_SET_MC_THREAD))
1704 kthread_run(lcs_register_mc_addresses, card, "regipm");
1712 lcs_get_control(struct lcs_card *card, struct lcs_cmd *cmd)
1719 lcs_schedule_recovery(card);
1722 if (card->dev) {
1724 card->dev->name);
1725 netif_carrier_off(card->dev);
1733 lcs_notify_lancmd_waiters(card, cmd);
1740 lcs_get_skb(struct lcs_card *card, char *skb_data, unsigned int skb_len)
1745 if (card->dev == NULL ||
1746 card->state != DEV_STATE_UP)
1747 /* The card isn't up. Ignore the packet. */
1752 dev_err(&card->dev->dev,
1754 card->dev->name);
1755 card->stats.rx_dropped++;
1759 skb->protocol = card->lan_type_trans(skb, card->dev);
1760 card->stats.rx_bytes += skb_len;
1761 card->stats.rx_packets++;
1763 *((__u32 *)skb->cb) = ++card->pkt_seq;
1773 struct lcs_card *card;
1783 card = container_of(channel, struct lcs_card, read);
1790 card->stats.rx_length_errors++;
1791 card->stats.rx_errors++;
1795 lcs_get_control(card, (struct lcs_cmd *) lcs_hdr);
1797 lcs_get_skb(card, (char *)(lcs_hdr + 1),
1801 dev_info_once(&card->dev->dev,
1809 lcs_ready_buffer(&card->read, buffer);
1818 struct lcs_card *card;
1821 card = (struct lcs_card *) dev->ml_priv;
1822 return &card->stats;
1832 struct lcs_card *card;
1836 card = (struct lcs_card *) dev->ml_priv;
1840 wait_event(card->write.wait_q,
1841 (card->write.state != LCS_CH_STATE_RUNNING));
1842 rc = lcs_stopcard(card);
1844 dev_err(&card->dev->dev,
1856 struct lcs_card *card;
1860 card = (struct lcs_card *) dev->ml_priv;
1862 rc = lcs_detect(card);
1870 card->state = DEV_STATE_UP;
1881 struct lcs_card *card;
1883 card = dev_get_drvdata(dev);
1885 if (!card)
1888 return sysfs_emit(buf, "%d\n", card->portno);
1897 struct lcs_card *card;
1901 card = dev_get_drvdata(dev);
1903 if (!card)
1910 card->portno = value;
1911 if (card->dev)
1912 card->dev->dev_port = card->portno;
1924 "OSA LCS card",
1947 struct lcs_card *card;
1949 card = dev_get_drvdata(dev);
1951 return card ? sysfs_emit(buf, "%u\n", card->lancmd_timeout) : 0;
1957 struct lcs_card *card;
1961 card = dev_get_drvdata(dev);
1963 if (!card)
1970 card->lancmd_timeout = value;
1982 struct lcs_card *card = dev_get_drvdata(dev);
1986 if (!card)
1988 if (card->state != DEV_STATE_UP)
1992 lcs_schedule_recovery(card);
2023 struct lcs_card *card;
2029 card = lcs_alloc_card();
2030 if (!card) {
2035 dev_set_drvdata(&ccwgdev->dev, card);
2038 card->gdev = ccwgdev;
2039 INIT_WORK(&card->kernel_thread_starter, lcs_start_kernel_thread);
2040 card->thread_start_mask = 0;
2041 card->thread_allowed_mask = 0;
2042 card->thread_running_mask = 0;
2051 struct lcs_card *card;
2054 card = dev_get_drvdata(&ccwgdev->dev);
2055 if (card->dev->reg_state != NETREG_UNINITIALIZED)
2057 SET_NETDEV_DEV(card->dev, &ccwgdev->dev);
2058 return register_netdev(card->dev);
2082 struct lcs_card *card;
2087 card = dev_get_drvdata(&ccwgdev->dev);
2088 if (!card)
2092 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2093 card->read.ccwdev = ccwgdev->cdev[0];
2094 card->write.ccwdev = ccwgdev->cdev[1];
2096 recover_state = card->state;
2097 rc = ccw_device_set_online(card->read.ccwdev);
2100 rc = ccw_device_set_online(card->write.ccwdev);
2106 lcs_setup_card(card);
2107 rc = lcs_detect(card);
2113 lcs_stopcard(card);
2116 if (card->dev) {
2118 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2121 switch (card->lan_type) {
2123 card->lan_type_trans = eth_type_trans;
2133 card->dev = dev;
2134 card->dev->ml_priv = card;
2135 card->dev->netdev_ops = &lcs_netdev_ops;
2136 card->dev->dev_port = card->portno;
2137 eth_hw_addr_set(card->dev, card->mac);
2139 if (!lcs_check_multicast_support(card))
2140 card->dev->netdev_ops = &lcs_mc_netdev_ops;
2143 lcs_set_allowed_threads(card,0xffffffff);
2145 lcs_set_multicast_list(card->dev);
2146 card->dev->flags |= IFF_UP;
2147 netif_carrier_on(card->dev);
2148 netif_wake_queue(card->dev);
2149 card->state = DEV_STATE_UP;
2151 lcs_stopcard(card);
2158 pr_info("LCS device %s %s IPv6 support\n", card->dev->name,
2159 (card->ip_assists_supported & LCS_IPASS_IPV6_SUPPORT) ?
2162 pr_info("LCS device %s %s Multicast support\n", card->dev->name,
2163 (card->ip_assists_supported & LCS_IPASS_MULTICAST_SUPPORT) ?
2168 ccw_device_set_offline(card->write.ccwdev);
2170 ccw_device_set_offline(card->read.ccwdev);
2181 struct lcs_card *card;
2186 card = dev_get_drvdata(&ccwgdev->dev);
2187 if (!card)
2190 lcs_set_allowed_threads(card, 0);
2191 if (lcs_wait_for_threads(card, LCS_SET_MC_THREAD))
2194 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2195 recover_state = card->state;
2197 ret = lcs_stop_device(card->dev);
2198 ret2 = ccw_device_set_offline(card->read.ccwdev);
2199 ret3 = ccw_device_set_offline(card->write.ccwdev);
2205 card->state = DEV_STATE_RECOVER;
2222 struct lcs_card *card;
2226 card = (struct lcs_card *) ptr;
2229 if (!lcs_do_run_thread(card, LCS_RECOVERY_THREAD))
2232 gdev = card->gdev;
2239 card->dev->name);
2242 card->dev->name);
2243 lcs_clear_thread_running_bit(card, LCS_RECOVERY_THREAD);
2248 * lcs_remove_device, free buffers and card
2253 struct lcs_card *card;
2255 card = dev_get_drvdata(&ccwgdev->dev);
2256 if (!card)
2260 LCS_DBF_HEX(3, setup, &card, sizeof(void*));
2264 if (card->dev)
2265 unregister_netdev(card->dev);
2266 lcs_cleanup_card(card);
2267 lcs_free_card(card);