Lines Matching refs:lynx

85 lynx_get(struct pcilynx *lynx)
87 kref_get(&lynx->kref);
89 return lynx;
99 lynx_put(struct pcilynx *lynx)
101 kref_put(&lynx->kref, lynx_release);
105 struct pcilynx *lynx;
145 list_empty(&client->lynx->link))
214 reg_write(struct pcilynx *lynx, int offset, u32 data)
216 writel(data, lynx->registers + offset);
220 reg_read(struct pcilynx *lynx, int offset)
222 return readl(lynx->registers + offset);
226 reg_set_bits(struct pcilynx *lynx, int offset, u32 mask)
228 reg_write(lynx, offset, (reg_read(lynx, offset) | mask));
236 run_pcl(struct pcilynx *lynx, dma_addr_t pcl_bus,
239 reg_write(lynx, DMA0_CURRENT_PCL + dmachan * 0x20, pcl_bus);
240 reg_write(lynx, DMA0_CHAN_CTRL + dmachan * 0x20,
245 set_phy_reg(struct pcilynx *lynx, int addr, int val)
248 dev_err(&lynx->pci_device->dev,
253 dev_err(&lynx->pci_device->dev,
257 reg_write(lynx, LINK_PHY, LINK_PHY_WRITE |
268 struct pcilynx *tmp, *lynx = NULL;
273 lynx = lynx_get(tmp);
277 if (lynx == NULL)
285 client->lynx = lynx;
296 lynx_put(lynx);
305 struct pcilynx *lynx = client->lynx;
307 spin_lock_irq(&lynx->client_list_lock);
309 spin_unlock_irq(&lynx->client_list_lock);
313 lynx_put(lynx);
329 if (list_empty(&client->lynx->link))
347 spinlock_t *client_list_lock = &client->lynx->client_list_lock;
367 list_add_tail(&client->link, &client->lynx->client_list);
406 packet_irq_handler(struct pcilynx *lynx)
415 length = __le32_to_cpu(lynx->rcv_pcl->pcl_status) & 0x00001fff;
416 tcode = __le32_to_cpu(lynx->rcv_buffer[1]) >> 4 & 0xf;
420 lynx->rcv_buffer[0] = (__force __le32)timestamp;
427 spin_lock(&lynx->client_list_lock);
429 list_for_each_entry(client, &lynx->client_list, link)
432 lynx->rcv_buffer, length + 4);
434 spin_unlock(&lynx->client_list_lock);
438 bus_reset_irq_handler(struct pcilynx *lynx)
447 spin_lock(&lynx->client_list_lock);
449 list_for_each_entry(client, &lynx->client_list, link)
452 spin_unlock(&lynx->client_list_lock);
458 struct pcilynx *lynx = device;
461 pci_int_status = reg_read(lynx, PCI_INT_STATUS);
474 link_int_status = reg_read(lynx, LINK_INT_STATUS);
475 reg_write(lynx, LINK_INT_STATUS, link_int_status);
478 bus_reset_irq_handler(lynx);
485 reg_write(lynx, PCI_INT_STATUS, pci_int_status);
488 packet_irq_handler(lynx);
489 run_pcl(lynx, lynx->rcv_start_pcl_bus, 0);
498 struct pcilynx *lynx = pci_get_drvdata(dev);
502 list_del_init(&lynx->link);
503 misc_deregister(&lynx->misc);
506 reg_write(lynx, PCI_INT_ENABLE, 0);
507 free_irq(lynx->pci_device->irq, lynx);
509 spin_lock_irq(&lynx->client_list_lock);
510 list_for_each_entry(client, &lynx->client_list, link)
512 spin_unlock_irq(&lynx->client_list_lock);
514 pci_free_consistent(lynx->pci_device, sizeof(struct pcl),
515 lynx->rcv_start_pcl, lynx->rcv_start_pcl_bus);
516 pci_free_consistent(lynx->pci_device, sizeof(struct pcl),
517 lynx->rcv_pcl, lynx->rcv_pcl_bus);
518 pci_free_consistent(lynx->pci_device, PAGE_SIZE,
519 lynx->rcv_buffer, lynx->rcv_buffer_bus);
521 iounmap(lynx->registers);
523 lynx_put(lynx);
531 struct pcilynx *lynx;
546 lynx = kzalloc(sizeof *lynx, GFP_KERNEL);
547 if (lynx == NULL) {
552 lynx->pci_device = dev;
553 pci_set_drvdata(dev, lynx);
555 spin_lock_init(&lynx->client_list_lock);
556 INIT_LIST_HEAD(&lynx->client_list);
557 kref_init(&lynx->kref);
559 lynx->registers = ioremap(pci_resource_start(dev, 0),
561 if (lynx->registers == NULL) {
567 lynx->rcv_start_pcl = pci_alloc_consistent(lynx->pci_device,
568 sizeof(struct pcl), &lynx->rcv_start_pcl_bus);
569 lynx->rcv_pcl = pci_alloc_consistent(lynx->pci_device,
570 sizeof(struct pcl), &lynx->rcv_pcl_bus);
571 lynx->rcv_buffer = pci_alloc_consistent(lynx->pci_device,
572 RCV_BUFFER_SIZE, &lynx->rcv_buffer_bus);
573 if (lynx->rcv_start_pcl == NULL ||
574 lynx->rcv_pcl == NULL ||
575 lynx->rcv_buffer == NULL) {
580 lynx->rcv_start_pcl->next = cpu_to_le32(lynx->rcv_pcl_bus);
581 lynx->rcv_pcl->next = cpu_to_le32(PCL_NEXT_INVALID);
582 lynx->rcv_pcl->async_error_next = cpu_to_le32(PCL_NEXT_INVALID);
584 lynx->rcv_pcl->buffer[0].control =
586 lynx->rcv_pcl->buffer[0].pointer =
587 cpu_to_le32(lynx->rcv_buffer_bus + 4);
588 p = lynx->rcv_buffer_bus + 2048;
589 end = lynx->rcv_buffer_bus + RCV_BUFFER_SIZE;
591 lynx->rcv_pcl->buffer[i].control =
593 lynx->rcv_pcl->buffer[i].pointer = cpu_to_le32(p);
595 lynx->rcv_pcl->buffer[i - 1].control |= cpu_to_le32(PCL_LAST_BUFF);
597 reg_set_bits(lynx, MISC_CONTROL, MISC_CONTROL_SWRESET);
599 reg_write(lynx, DMA0_CHAN_CTRL, 0);
600 reg_write(lynx, DMA_GLOBAL_REGISTER, 0x00 << 24);
604 if ((get_phy_reg(lynx, 2) & 0xe0) == 0xe0) {
605 lynx->phyic.reg_1394a = 1;
606 PRINT(KERN_INFO, lynx->id,
608 lynx->phyic.vendor = get_phy_vendorid(lynx);
609 lynx->phyic.product = get_phy_productid(lynx);
611 lynx->phyic.reg_1394a = 0;
612 PRINT(KERN_INFO, lynx->id, "found old 1394 PHY");
617 reg_write(lynx, FIFO_SIZES, 255);
619 reg_set_bits(lynx, PCI_INT_ENABLE, PCI_INT_DMA_ALL);
621 reg_write(lynx, LINK_INT_ENABLE,
629 set_phy_reg(lynx, 4, 0);
632 reg_set_bits(lynx, LINK_CONTROL, LINK_CONTROL_SNOOP_ENABLE);
634 run_pcl(lynx, lynx->rcv_start_pcl_bus, 0);
637 driver_name, lynx)) {
644 lynx->misc.parent = &dev->dev;
645 lynx->misc.minor = MISC_DYNAMIC_MINOR;
646 lynx->misc.name = "nosy";
647 lynx->misc.fops = &nosy_ops;
650 ret = misc_register(&lynx->misc);
656 list_add_tail(&lynx->link, &card_list);
665 reg_write(lynx, PCI_INT_ENABLE, 0);
666 free_irq(lynx->pci_device->irq, lynx);
669 if (lynx->rcv_start_pcl)
670 pci_free_consistent(lynx->pci_device, sizeof(struct pcl),
671 lynx->rcv_start_pcl, lynx->rcv_start_pcl_bus);
672 if (lynx->rcv_pcl)
673 pci_free_consistent(lynx->pci_device, sizeof(struct pcl),
674 lynx->rcv_pcl, lynx->rcv_pcl_bus);
675 if (lynx->rcv_buffer)
676 pci_free_consistent(lynx->pci_device, PAGE_SIZE,
677 lynx->rcv_buffer, lynx->rcv_buffer_bus);
678 iounmap(lynx->registers);
681 kfree(lynx);