Lines Matching refs:he_dev

88 #define hprintk(fmt,args...)	printk(KERN_ERR DEV_LABEL "%d: " fmt, he_dev->number , ##args)
91 #define HPRINTK(fmt,args...) printk(KERN_DEBUG DEV_LABEL "%d: " fmt, he_dev->number , ##args)
106 static void he_stop(struct he_dev *dev);
110 static u8 read_prom_byte(struct he_dev *he_dev, int addr);
114 static struct he_dev *he_devs;
182 he_writel_internal(struct he_dev *he_dev, unsigned val, unsigned addr,
185 he_writel(he_dev, val, CON_DAT);
186 (void) he_readl(he_dev, CON_DAT); /* flush posted writes */
187 he_writel(he_dev, flags | CON_CTL_WRITE | CON_CTL_ADDR(addr), CON_CTL);
188 while (he_readl(he_dev, CON_CTL) & CON_CTL_BUSY);
201 he_readl_internal(struct he_dev *he_dev, unsigned addr, unsigned flags)
203 he_writel(he_dev, flags | CON_CTL_READ | CON_CTL_ADDR(addr), CON_CTL);
204 while (he_readl(he_dev, CON_CTL) & CON_CTL_BUSY);
205 return he_readl(he_dev, CON_DAT);
328 __find_vcc(struct he_dev *he_dev, unsigned cid)
336 vpi = cid >> he_dev->vcibits;
337 vci = cid & ((1 << he_dev->vcibits) - 1);
342 if (vcc->dev == he_dev->atm_dev &&
355 struct he_dev *he_dev = NULL;
375 he_dev = kzalloc(sizeof(struct he_dev),
377 if (!he_dev) {
381 he_dev->pci_dev = pci_dev;
382 he_dev->atm_dev = atm_dev;
383 he_dev->atm_dev->dev_data = he_dev;
384 atm_dev->dev_data = he_dev;
385 he_dev->number = atm_dev->number;
386 tasklet_init(&he_dev->tasklet, he_tasklet, (unsigned long) he_dev);
387 spin_lock_init(&he_dev->global_lock);
390 he_stop(he_dev);
394 he_dev->next = NULL;
396 he_dev->next = he_devs;
397 he_devs = he_dev;
403 kfree(he_dev);
411 struct he_dev *he_dev;
414 he_dev = HE_DEV(atm_dev);
418 he_stop(he_dev);
420 kfree(he_dev);
445 static void he_init_rx_lbfp0(struct he_dev *he_dev)
448 unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
449 unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
450 unsigned row_offset = he_dev->r0_startrow * he_dev->bytes_per_row;
453 lbm_offset = he_readl(he_dev, RCMLBM_BA);
455 he_writel(he_dev, lbufd_index, RLBF0_H);
457 for (i = 0, lbuf_count = 0; i < he_dev->r0_numbuffs; ++i) {
461 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
462 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
466 row_offset += he_dev->bytes_per_row;
471 he_writel(he_dev, lbufd_index - 2, RLBF0_T);
472 he_writel(he_dev, he_dev->r0_numbuffs, RLBF0_C);
475 static void he_init_rx_lbfp1(struct he_dev *he_dev)
478 unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
479 unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
480 unsigned row_offset = he_dev->r1_startrow * he_dev->bytes_per_row;
483 lbm_offset = he_readl(he_dev, RCMLBM_BA) + (2 * lbufd_index);
485 he_writel(he_dev, lbufd_index, RLBF1_H);
487 for (i = 0, lbuf_count = 0; i < he_dev->r1_numbuffs; ++i) {
491 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
492 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
496 row_offset += he_dev->bytes_per_row;
501 he_writel(he_dev, lbufd_index - 2, RLBF1_T);
502 he_writel(he_dev, he_dev->r1_numbuffs, RLBF1_C);
505 static void he_init_tx_lbfp(struct he_dev *he_dev)
508 unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
509 unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
510 unsigned row_offset = he_dev->tx_startrow * he_dev->bytes_per_row;
512 lbufd_index = he_dev->r0_numbuffs + he_dev->r1_numbuffs;
513 lbm_offset = he_readl(he_dev, RCMLBM_BA) + (2 * lbufd_index);
515 he_writel(he_dev, lbufd_index, TLBF_H);
517 for (i = 0, lbuf_count = 0; i < he_dev->tx_numbuffs; ++i) {
521 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
522 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
526 row_offset += he_dev->bytes_per_row;
531 he_writel(he_dev, lbufd_index - 1, TLBF_T);
534 static int he_init_tpdrq(struct he_dev *he_dev)
536 he_dev->tpdrq_base = dma_alloc_coherent(&he_dev->pci_dev->dev,
538 &he_dev->tpdrq_phys,
540 if (he_dev->tpdrq_base == NULL) {
545 he_dev->tpdrq_tail = he_dev->tpdrq_base;
546 he_dev->tpdrq_head = he_dev->tpdrq_base;
548 he_writel(he_dev, he_dev->tpdrq_phys, TPDRQ_B_H);
549 he_writel(he_dev, 0, TPDRQ_T);
550 he_writel(he_dev, CONFIG_TPDRQ_SIZE - 1, TPDRQ_S);
555 static void he_init_cs_block(struct he_dev *he_dev)
563 he_writel_mbox(he_dev, 0x0, CS_STTIM0 + reg);
567 clock = he_is622(he_dev) ? 66667000 : 50000000;
568 rate = he_dev->atm_dev->link_rate;
579 he_writel_mbox(he_dev, period, CS_TGRLD0 + reg);
583 if (he_is622(he_dev)) {
585 he_writel_mbox(he_dev, 0x000800fa, CS_ERTHR0);
586 he_writel_mbox(he_dev, 0x000c33cb, CS_ERTHR1);
587 he_writel_mbox(he_dev, 0x0010101b, CS_ERTHR2);
588 he_writel_mbox(he_dev, 0x00181dac, CS_ERTHR3);
589 he_writel_mbox(he_dev, 0x00280600, CS_ERTHR4);
592 he_writel_mbox(he_dev, 0x023de8b3, CS_ERCTL0);
593 he_writel_mbox(he_dev, 0x1801, CS_ERCTL1);
594 he_writel_mbox(he_dev, 0x68b3, CS_ERCTL2);
595 he_writel_mbox(he_dev, 0x1280, CS_ERSTAT0);
596 he_writel_mbox(he_dev, 0x68b3, CS_ERSTAT1);
597 he_writel_mbox(he_dev, 0x14585, CS_RTFWR);
599 he_writel_mbox(he_dev, 0x4680, CS_RTATR);
602 he_writel_mbox(he_dev, 0x00159ece, CS_TFBSET);
603 he_writel_mbox(he_dev, 0x68b3, CS_WCRMAX);
604 he_writel_mbox(he_dev, 0x5eb3, CS_WCRMIN);
605 he_writel_mbox(he_dev, 0xe8b3, CS_WCRINC);
606 he_writel_mbox(he_dev, 0xdeb3, CS_WCRDEC);
607 he_writel_mbox(he_dev, 0x68b3, CS_WCRCEIL);
610 he_writel_mbox(he_dev, 0x5, CS_OTPPER);
611 he_writel_mbox(he_dev, 0x14, CS_OTWPER);
614 he_writel_mbox(he_dev, 0x000400ea, CS_ERTHR0);
615 he_writel_mbox(he_dev, 0x00063388, CS_ERTHR1);
616 he_writel_mbox(he_dev, 0x00081018, CS_ERTHR2);
617 he_writel_mbox(he_dev, 0x000c1dac, CS_ERTHR3);
618 he_writel_mbox(he_dev, 0x0014051a, CS_ERTHR4);
621 he_writel_mbox(he_dev, 0x0235e4b1, CS_ERCTL0);
622 he_writel_mbox(he_dev, 0x4701, CS_ERCTL1);
623 he_writel_mbox(he_dev, 0x64b1, CS_ERCTL2);
624 he_writel_mbox(he_dev, 0x1280, CS_ERSTAT0);
625 he_writel_mbox(he_dev, 0x64b1, CS_ERSTAT1);
626 he_writel_mbox(he_dev, 0xf424, CS_RTFWR);
628 he_writel_mbox(he_dev, 0x4680, CS_RTATR);
631 he_writel_mbox(he_dev, 0x000563b7, CS_TFBSET);
632 he_writel_mbox(he_dev, 0x64b1, CS_WCRMAX);
633 he_writel_mbox(he_dev, 0x5ab1, CS_WCRMIN);
634 he_writel_mbox(he_dev, 0xe4b1, CS_WCRINC);
635 he_writel_mbox(he_dev, 0xdab1, CS_WCRDEC);
636 he_writel_mbox(he_dev, 0x64b1, CS_WCRCEIL);
639 he_writel_mbox(he_dev, 0x6, CS_OTPPER);
640 he_writel_mbox(he_dev, 0x1e, CS_OTWPER);
643 he_writel_mbox(he_dev, 0x8, CS_OTTLIM);
646 he_writel_mbox(he_dev, 0x0, CS_HGRRT0 + reg);
650 static int he_init_cs_block_rcm(struct he_dev *he_dev)
667 he_writel_rcm(he_dev, 0x0, CONFIG_RCMABR + reg);
672 he_writel_rcm(he_dev, 0x0, CONFIG_RCMABR + reg);
682 rate = he_dev->atm_dev->link_rate;
737 buf = rate_cps * he_dev->tx_numbuffs /
738 (he_dev->atm_dev->link_rate * 2);
741 mult = he_dev->atm_dev->link_rate / ATM_OC3_PCR;
760 he_writel_rcm(he_dev, reg,
770 static int he_init_group(struct he_dev *he_dev, int group)
776 he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32));
777 he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32));
778 he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32));
779 he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
783 he_dev->rbpl_table = bitmap_zalloc(RBPL_TABLE_SIZE, GFP_KERNEL);
784 if (!he_dev->rbpl_table) {
790 he_dev->rbpl_virt = kmalloc_array(RBPL_TABLE_SIZE,
791 sizeof(*he_dev->rbpl_virt),
793 if (!he_dev->rbpl_virt) {
799 he_dev->rbpl_pool = dma_pool_create("rbpl", &he_dev->pci_dev->dev,
801 if (he_dev->rbpl_pool == NULL) {
806 he_dev->rbpl_base = dma_alloc_coherent(&he_dev->pci_dev->dev,
808 &he_dev->rbpl_phys, GFP_KERNEL);
809 if (he_dev->rbpl_base == NULL) {
814 INIT_LIST_HEAD(&he_dev->rbpl_outstanding);
818 heb = dma_pool_alloc(he_dev->rbpl_pool, GFP_KERNEL, &mapping);
822 list_add(&heb->entry, &he_dev->rbpl_outstanding);
824 set_bit(i, he_dev->rbpl_table);
825 he_dev->rbpl_virt[i] = heb;
826 he_dev->rbpl_hint = i + 1;
827 he_dev->rbpl_base[i].idx = i << RBP_IDX_OFFSET;
828 he_dev->rbpl_base[i].phys = mapping + offsetof(struct he_buff, data);
830 he_dev->rbpl_tail = &he_dev->rbpl_base[CONFIG_RBPL_SIZE - 1];
832 he_writel(he_dev, he_dev->rbpl_phys, G0_RBPL_S + (group * 32));
833 he_writel(he_dev, RBPL_MASK(he_dev->rbpl_tail),
835 he_writel(he_dev, (CONFIG_RBPL_BUFSIZE - sizeof(struct he_buff))/4,
837 he_writel(he_dev,
845 he_dev->rbrq_base = dma_alloc_coherent(&he_dev->pci_dev->dev,
847 &he_dev->rbrq_phys, GFP_KERNEL);
848 if (he_dev->rbrq_base == NULL) {
853 he_dev->rbrq_head = he_dev->rbrq_base;
854 he_writel(he_dev, he_dev->rbrq_phys, G0_RBRQ_ST + (group * 16));
855 he_writel(he_dev, 0, G0_RBRQ_H + (group * 16));
856 he_writel(he_dev,
861 he_writel(he_dev, RBRQ_TIME(768) | RBRQ_COUNT(7),
864 he_writel(he_dev, RBRQ_TIME(0) | RBRQ_COUNT(1),
869 he_dev->tbrq_base = dma_alloc_coherent(&he_dev->pci_dev->dev,
871 &he_dev->tbrq_phys, GFP_KERNEL);
872 if (he_dev->tbrq_base == NULL) {
877 he_dev->tbrq_head = he_dev->tbrq_base;
879 he_writel(he_dev, he_dev->tbrq_phys, G0_TBRQ_B_T + (group * 16));
880 he_writel(he_dev, 0, G0_TBRQ_H + (group * 16));
881 he_writel(he_dev, CONFIG_TBRQ_SIZE - 1, G0_TBRQ_S + (group * 16));
882 he_writel(he_dev, CONFIG_TBRQ_THRESH, G0_TBRQ_THRESH + (group * 16));
887 dma_free_coherent(&he_dev->pci_dev->dev, CONFIG_RBRQ_SIZE *
888 sizeof(struct he_rbrq), he_dev->rbrq_base,
889 he_dev->rbrq_phys);
891 list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
892 dma_pool_free(he_dev->rbpl_pool, heb, heb->mapping);
894 dma_free_coherent(&he_dev->pci_dev->dev, CONFIG_RBPL_SIZE *
895 sizeof(struct he_rbp), he_dev->rbpl_base,
896 he_dev->rbpl_phys);
898 dma_pool_destroy(he_dev->rbpl_pool);
900 kfree(he_dev->rbpl_virt);
902 bitmap_free(he_dev->rbpl_table);
907 static int he_init_irq(struct he_dev *he_dev)
914 he_dev->irq_base = dma_alloc_coherent(&he_dev->pci_dev->dev,
916 &he_dev->irq_phys, GFP_KERNEL);
917 if (he_dev->irq_base == NULL) {
921 he_dev->irq_tailoffset = (unsigned *)
922 &he_dev->irq_base[CONFIG_IRQ_SIZE];
923 *he_dev->irq_tailoffset = 0;
924 he_dev->irq_head = he_dev->irq_base;
925 he_dev->irq_tail = he_dev->irq_base;
928 he_dev->irq_base[i].isw = ITYPE_INVALID;
930 he_writel(he_dev, he_dev->irq_phys, IRQ0_BASE);
931 he_writel(he_dev,
934 he_writel(he_dev, IRQ_INT_A | IRQ_TYPE_LINE, IRQ0_CNTL);
935 he_writel(he_dev, 0x0, IRQ0_DATA);
937 he_writel(he_dev, 0x0, IRQ1_BASE);
938 he_writel(he_dev, 0x0, IRQ1_HEAD);
939 he_writel(he_dev, 0x0, IRQ1_CNTL);
940 he_writel(he_dev, 0x0, IRQ1_DATA);
942 he_writel(he_dev, 0x0, IRQ2_BASE);
943 he_writel(he_dev, 0x0, IRQ2_HEAD);
944 he_writel(he_dev, 0x0, IRQ2_CNTL);
945 he_writel(he_dev, 0x0, IRQ2_DATA);
947 he_writel(he_dev, 0x0, IRQ3_BASE);
948 he_writel(he_dev, 0x0, IRQ3_HEAD);
949 he_writel(he_dev, 0x0, IRQ3_CNTL);
950 he_writel(he_dev, 0x0, IRQ3_DATA);
954 he_writel(he_dev, 0x0, GRP_10_MAP);
955 he_writel(he_dev, 0x0, GRP_32_MAP);
956 he_writel(he_dev, 0x0, GRP_54_MAP);
957 he_writel(he_dev, 0x0, GRP_76_MAP);
959 if (request_irq(he_dev->pci_dev->irq,
960 he_irq_handler, IRQF_SHARED, DEV_LABEL, he_dev)) {
961 hprintk("irq %d already in use\n", he_dev->pci_dev->irq);
965 he_dev->irq = he_dev->pci_dev->irq;
972 struct he_dev *he_dev;
984 he_dev = HE_DEV(dev);
985 pci_dev = he_dev->pci_dev;
1048 if (!(he_dev->membase = ioremap(membase, HE_REGMAP_SIZE))) {
1054 he_writel(he_dev, 0x0, RESET_CNTL);
1055 he_writel(he_dev, 0xff, RESET_CNTL);
1058 status = he_readl(he_dev, RESET_CNTL);
1065 host_cntl = he_readl(he_dev, HOST_CNTL);
1083 he_dev->prod_id[i] = read_prom_byte(he_dev, PROD_ID + i);
1085 he_dev->media = read_prom_byte(he_dev, MEDIA);
1088 dev->esi[i] = read_prom_byte(he_dev, MAC_ADDR + i);
1090 hprintk("%s%s, %pM\n", he_dev->prod_id,
1091 he_dev->media & 0x40 ? "SM" : "MM", dev->esi);
1092 he_dev->atm_dev->link_rate = he_is622(he_dev) ?
1096 lb_swap = he_readl(he_dev, LB_SWAP);
1097 if (he_is622(he_dev))
1107 he_writel(he_dev, lb_swap, LB_SWAP);
1110 he_writel(he_dev, he_is622(he_dev) ? LB_64_ENB : 0x0, SDRAM_CTL);
1114 he_writel(he_dev, lb_swap, LB_SWAP);
1117 if ((err = he_init_irq(he_dev)) != 0)
1123 he_writel(he_dev, host_cntl, HOST_CNTL);
1163 he_dev->vcibits = CONFIG_DEFAULT_VCIBITS;
1164 he_dev->vpibits = CONFIG_DEFAULT_VPIBITS;
1172 he_dev->vpibits = nvpibits;
1173 he_dev->vcibits = HE_MAXCIDBITS - nvpibits;
1177 he_dev->vcibits = nvcibits;
1178 he_dev->vpibits = HE_MAXCIDBITS - nvcibits;
1182 if (he_is622(he_dev)) {
1183 he_dev->cells_per_row = 40;
1184 he_dev->bytes_per_row = 2048;
1185 he_dev->r0_numrows = 256;
1186 he_dev->tx_numrows = 512;
1187 he_dev->r1_numrows = 256;
1188 he_dev->r0_startrow = 0;
1189 he_dev->tx_startrow = 256;
1190 he_dev->r1_startrow = 768;
1192 he_dev->cells_per_row = 20;
1193 he_dev->bytes_per_row = 1024;
1194 he_dev->r0_numrows = 512;
1195 he_dev->tx_numrows = 1018;
1196 he_dev->r1_numrows = 512;
1197 he_dev->r0_startrow = 6;
1198 he_dev->tx_startrow = 518;
1199 he_dev->r1_startrow = 1536;
1202 he_dev->cells_per_lbuf = 4;
1203 he_dev->buffer_limit = 4;
1204 he_dev->r0_numbuffs = he_dev->r0_numrows *
1205 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1206 if (he_dev->r0_numbuffs > 2560)
1207 he_dev->r0_numbuffs = 2560;
1209 he_dev->r1_numbuffs = he_dev->r1_numrows *
1210 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1211 if (he_dev->r1_numbuffs > 2560)
1212 he_dev->r1_numbuffs = 2560;
1214 he_dev->tx_numbuffs = he_dev->tx_numrows *
1215 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1216 if (he_dev->tx_numbuffs > 5120)
1217 he_dev->tx_numbuffs = 5120;
1221 he_writel(he_dev,
1224 (he_is622(he_dev) ? BUS_MULTI(0x28) : BUS_MULTI(0x46)) |
1225 (he_is622(he_dev) ? NET_PREF(0x50) : NET_PREF(0x8c)),
1228 he_writel(he_dev, BANK_ON |
1229 (he_is622(he_dev) ? (REF_RATE(0x384) | WIDE_DATA) : REF_RATE(0x150)),
1232 he_writel(he_dev,
1233 (he_is622(he_dev) ? RM_BANK_WAIT(1) : RM_BANK_WAIT(0)) |
1235 he_writel(he_dev,
1236 (he_is622(he_dev) ? TM_BANK_WAIT(2) : TM_BANK_WAIT(1)) |
1239 he_writel(he_dev, he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD, LB_CONFIG);
1241 he_writel(he_dev,
1242 (he_is622(he_dev) ? UT_RD_DELAY(8) : UT_RD_DELAY(0)) |
1243 (he_is622(he_dev) ? RC_UT_MODE(0) : RC_UT_MODE(1)) |
1244 RX_VALVP(he_dev->vpibits) |
1245 RX_VALVC(he_dev->vcibits), RC_CONFIG);
1247 he_writel(he_dev, DRF_THRESH(0x20) |
1248 (he_is622(he_dev) ? TX_UT_MODE(0) : TX_UT_MODE(1)) |
1249 TX_VCI_MASK(he_dev->vcibits) |
1250 LBFREE_CNT(he_dev->tx_numbuffs), TX_CONFIG);
1252 he_writel(he_dev, 0x0, TXAAL5_PROTO);
1254 he_writel(he_dev, PHY_INT_ENB |
1255 (he_is622(he_dev) ? PTMR_PRE(67 - 1) : PTMR_PRE(50 - 1)),
1261 he_writel_tcm(he_dev, 0, i);
1264 he_writel_rcm(he_dev, 0, i);
1297 he_writel(he_dev, CONFIG_TSRB, TSRB_BA);
1298 he_writel(he_dev, CONFIG_TSRC, TSRC_BA);
1299 he_writel(he_dev, CONFIG_TSRD, TSRD_BA);
1300 he_writel(he_dev, CONFIG_TMABR, TMABR_BA);
1301 he_writel(he_dev, CONFIG_TPDBA, TPD_BA);
1331 he_writel(he_dev, 0x08000, RCMLBM_BA);
1332 he_writel(he_dev, 0x0e000, RCMRSRB_BA);
1333 he_writel(he_dev, 0x0d800, RCMABR_BA);
1337 he_init_rx_lbfp0(he_dev);
1338 he_init_rx_lbfp1(he_dev);
1340 he_writel(he_dev, 0x0, RLBC_H);
1341 he_writel(he_dev, 0x0, RLBC_T);
1342 he_writel(he_dev, 0x0, RLBC_H2);
1344 he_writel(he_dev, 512, RXTHRSH); /* 10% of r0+r1 buffers */
1345 he_writel(he_dev, 256, LITHRSH); /* 5% of r0+r1 buffers */
1347 he_init_tx_lbfp(he_dev);
1349 he_writel(he_dev, he_is622(he_dev) ? 0x104780 : 0x800, UBUFF_BA);
1353 if (he_is622(he_dev)) {
1354 he_writel(he_dev, 0x000f, G0_INMQ_S);
1355 he_writel(he_dev, 0x200f, G0_INMQ_L);
1357 he_writel(he_dev, 0x001f, G1_INMQ_S);
1358 he_writel(he_dev, 0x201f, G1_INMQ_L);
1360 he_writel(he_dev, 0x002f, G2_INMQ_S);
1361 he_writel(he_dev, 0x202f, G2_INMQ_L);
1363 he_writel(he_dev, 0x003f, G3_INMQ_S);
1364 he_writel(he_dev, 0x203f, G3_INMQ_L);
1366 he_writel(he_dev, 0x004f, G4_INMQ_S);
1367 he_writel(he_dev, 0x204f, G4_INMQ_L);
1369 he_writel(he_dev, 0x005f, G5_INMQ_S);
1370 he_writel(he_dev, 0x205f, G5_INMQ_L);
1372 he_writel(he_dev, 0x006f, G6_INMQ_S);
1373 he_writel(he_dev, 0x206f, G6_INMQ_L);
1375 he_writel(he_dev, 0x007f, G7_INMQ_S);
1376 he_writel(he_dev, 0x207f, G7_INMQ_L);
1378 he_writel(he_dev, 0x0000, G0_INMQ_S);
1379 he_writel(he_dev, 0x0008, G0_INMQ_L);
1381 he_writel(he_dev, 0x0001, G1_INMQ_S);
1382 he_writel(he_dev, 0x0009, G1_INMQ_L);
1384 he_writel(he_dev, 0x0002, G2_INMQ_S);
1385 he_writel(he_dev, 0x000a, G2_INMQ_L);
1387 he_writel(he_dev, 0x0003, G3_INMQ_S);
1388 he_writel(he_dev, 0x000b, G3_INMQ_L);
1390 he_writel(he_dev, 0x0004, G4_INMQ_S);
1391 he_writel(he_dev, 0x000c, G4_INMQ_L);
1393 he_writel(he_dev, 0x0005, G5_INMQ_S);
1394 he_writel(he_dev, 0x000d, G5_INMQ_L);
1396 he_writel(he_dev, 0x0006, G6_INMQ_S);
1397 he_writel(he_dev, 0x000e, G6_INMQ_L);
1399 he_writel(he_dev, 0x0007, G7_INMQ_S);
1400 he_writel(he_dev, 0x000f, G7_INMQ_L);
1405 he_writel(he_dev, 0x0, MCC);
1406 he_writel(he_dev, 0x0, OEC);
1407 he_writel(he_dev, 0x0, DCC);
1408 he_writel(he_dev, 0x0, CEC);
1412 he_init_cs_block(he_dev);
1416 if (he_init_cs_block_rcm(he_dev) < 0)
1421 he_init_tpdrq(he_dev);
1423 he_dev->tpd_pool = dma_pool_create("tpd", &he_dev->pci_dev->dev,
1425 if (he_dev->tpd_pool == NULL) {
1430 INIT_LIST_HEAD(&he_dev->outstanding_tpds);
1432 if (he_init_group(he_dev, 0) != 0)
1436 he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32));
1437 he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32));
1438 he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32));
1439 he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
1442 he_writel(he_dev, 0x0, G0_RBPL_S + (group * 32));
1443 he_writel(he_dev, 0x0, G0_RBPL_T + (group * 32));
1444 he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
1446 he_writel(he_dev, 0x0, G0_RBPL_BS + (group * 32));
1448 he_writel(he_dev, 0x0, G0_RBRQ_ST + (group * 16));
1449 he_writel(he_dev, 0x0, G0_RBRQ_H + (group * 16));
1450 he_writel(he_dev, RBRQ_THRESH(0x1) | RBRQ_SIZE(0x0),
1452 he_writel(he_dev, 0x0, G0_RBRQ_I + (group * 16));
1454 he_writel(he_dev, 0x0, G0_TBRQ_B_T + (group * 16));
1455 he_writel(he_dev, 0x0, G0_TBRQ_H + (group * 16));
1456 he_writel(he_dev, TBRQ_THRESH(0x1),
1458 he_writel(he_dev, 0x0, G0_TBRQ_S + (group * 16));
1463 he_dev->hsp = dma_alloc_coherent(&he_dev->pci_dev->dev,
1465 &he_dev->hsp_phys, GFP_KERNEL);
1466 if (he_dev->hsp == NULL) {
1470 he_writel(he_dev, he_dev->hsp_phys, HSP_BA);
1475 if (he_isMM(he_dev))
1476 suni_init(he_dev->atm_dev);
1477 if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->start)
1478 he_dev->atm_dev->phy->start(he_dev->atm_dev);
1485 val = he_phy_get(he_dev->atm_dev, SUNI_TPOP_APM);
1487 he_phy_put(he_dev->atm_dev, val, SUNI_TPOP_APM);
1488 he_phy_put(he_dev->atm_dev, SUNI_TACP_IUCHP_CLP, SUNI_TACP_IUCHP);
1493 reg = he_readl_mbox(he_dev, CS_ERCTL0);
1495 he_writel_mbox(he_dev, reg, CS_ERCTL0);
1497 reg = he_readl(he_dev, RC_CONFIG);
1499 he_writel(he_dev, reg, RC_CONFIG);
1502 he_dev->cs_stper[i].inuse = 0;
1503 he_dev->cs_stper[i].pcr = -1;
1505 he_dev->total_bw = 0;
1510 he_dev->atm_dev->ci_range.vpi_bits = he_dev->vpibits;
1511 he_dev->atm_dev->ci_range.vci_bits = he_dev->vcibits;
1513 he_dev->irq_peak = 0;
1514 he_dev->rbrq_peak = 0;
1515 he_dev->rbpl_peak = 0;
1516 he_dev->tbrq_peak = 0;
1524 he_stop(struct he_dev *he_dev)
1531 pci_dev = he_dev->pci_dev;
1535 if (he_dev->membase) {
1540 tasklet_disable(&he_dev->tasklet);
1544 reg = he_readl_mbox(he_dev, CS_ERCTL0);
1546 he_writel_mbox(he_dev, reg, CS_ERCTL0);
1548 reg = he_readl(he_dev, RC_CONFIG);
1550 he_writel(he_dev, reg, RC_CONFIG);
1554 if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->stop)
1555 he_dev->atm_dev->phy->stop(he_dev->atm_dev);
1558 if (he_dev->irq)
1559 free_irq(he_dev->irq, he_dev);
1561 if (he_dev->irq_base)
1562 dma_free_coherent(&he_dev->pci_dev->dev, (CONFIG_IRQ_SIZE + 1)
1563 * sizeof(struct he_irq), he_dev->irq_base, he_dev->irq_phys);
1565 if (he_dev->hsp)
1566 dma_free_coherent(&he_dev->pci_dev->dev, sizeof(struct he_hsp),
1567 he_dev->hsp, he_dev->hsp_phys);
1569 if (he_dev->rbpl_base) {
1570 list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
1571 dma_pool_free(he_dev->rbpl_pool, heb, heb->mapping);
1573 dma_free_coherent(&he_dev->pci_dev->dev, CONFIG_RBPL_SIZE
1574 * sizeof(struct he_rbp), he_dev->rbpl_base, he_dev->rbpl_phys);
1577 kfree(he_dev->rbpl_virt);
1578 bitmap_free(he_dev->rbpl_table);
1579 dma_pool_destroy(he_dev->rbpl_pool);
1581 if (he_dev->rbrq_base)
1582 dma_free_coherent(&he_dev->pci_dev->dev, CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq),
1583 he_dev->rbrq_base, he_dev->rbrq_phys);
1585 if (he_dev->tbrq_base)
1586 dma_free_coherent(&he_dev->pci_dev->dev, CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq),
1587 he_dev->tbrq_base, he_dev->tbrq_phys);
1589 if (he_dev->tpdrq_base)
1590 dma_free_coherent(&he_dev->pci_dev->dev, CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq),
1591 he_dev->tpdrq_base, he_dev->tpdrq_phys);
1593 dma_pool_destroy(he_dev->tpd_pool);
1595 if (he_dev->pci_dev) {
1596 pci_read_config_word(he_dev->pci_dev, PCI_COMMAND, &command);
1598 pci_write_config_word(he_dev->pci_dev, PCI_COMMAND, command);
1601 if (he_dev->membase)
1602 iounmap(he_dev->membase);
1606 __alloc_tpd(struct he_dev *he_dev)
1611 tpd = dma_pool_alloc(he_dev->tpd_pool, GFP_ATOMIC, &mapping);
1639 he_service_rbrq(struct he_dev *he_dev, int group)
1642 ((unsigned long)he_dev->rbrq_base |
1643 he_dev->hsp->group[group].rbrq_tail);
1654 while (he_dev->rbrq_head != rbrq_tail) {
1658 he_dev->rbrq_head, group,
1659 RBRQ_ADDR(he_dev->rbrq_head),
1660 RBRQ_BUFLEN(he_dev->rbrq_head),
1661 RBRQ_CID(he_dev->rbrq_head),
1662 RBRQ_CRC_ERR(he_dev->rbrq_head) ? " CRC_ERR" : "",
1663 RBRQ_LEN_ERR(he_dev->rbrq_head) ? " LEN_ERR" : "",
1664 RBRQ_END_PDU(he_dev->rbrq_head) ? " END_PDU" : "",
1665 RBRQ_AAL5_PROT(he_dev->rbrq_head) ? " AAL5_PROT" : "",
1666 RBRQ_CON_CLOSED(he_dev->rbrq_head) ? " CON_CLOSED" : "",
1667 RBRQ_HBUF_ERR(he_dev->rbrq_head) ? " HBUF_ERR" : "");
1669 i = RBRQ_ADDR(he_dev->rbrq_head) >> RBP_IDX_OFFSET;
1670 heb = he_dev->rbpl_virt[i];
1672 cid = RBRQ_CID(he_dev->rbrq_head);
1674 vcc = __find_vcc(he_dev, cid);
1679 if (!RBRQ_HBUF_ERR(he_dev->rbrq_head)) {
1680 clear_bit(i, he_dev->rbpl_table);
1682 dma_pool_free(he_dev->rbpl_pool, heb, heb->mapping);
1688 if (RBRQ_HBUF_ERR(he_dev->rbrq_head)) {
1694 heb->len = RBRQ_BUFLEN(he_dev->rbrq_head) * 4;
1695 clear_bit(i, he_dev->rbpl_table);
1699 if (RBRQ_CON_CLOSED(he_dev->rbrq_head)) {
1706 if (!RBRQ_END_PDU(he_dev->rbrq_head))
1709 if (RBRQ_LEN_ERR(he_dev->rbrq_head)
1710 || RBRQ_CRC_ERR(he_dev->rbrq_head)) {
1712 RBRQ_CRC_ERR(he_dev->rbrq_head)
1714 RBRQ_LEN_ERR(he_dev->rbrq_head)
1765 spin_unlock(&he_dev->global_lock);
1767 spin_lock(&he_dev->global_lock);
1775 dma_pool_free(he_dev->rbpl_pool, heb, heb->mapping);
1780 he_dev->rbrq_head = (struct he_rbrq *)
1781 ((unsigned long) he_dev->rbrq_base |
1782 RBRQ_MASK(he_dev->rbrq_head + 1));
1788 if (updated > he_dev->rbrq_peak)
1789 he_dev->rbrq_peak = updated;
1791 he_writel(he_dev, RBRQ_MASK(he_dev->rbrq_head),
1799 he_service_tbrq(struct he_dev *he_dev, int group)
1802 ((unsigned long)he_dev->tbrq_base |
1803 he_dev->hsp->group[group].tbrq_tail);
1810 while (he_dev->tbrq_head != tbrq_tail) {
1815 TBRQ_TPD(he_dev->tbrq_head),
1816 TBRQ_EOS(he_dev->tbrq_head) ? " EOS" : "",
1817 TBRQ_MULTIPLE(he_dev->tbrq_head) ? " MULTIPLE" : "");
1819 list_for_each_entry(__tpd, &he_dev->outstanding_tpds, entry) {
1820 if (TPD_ADDR(__tpd->status) == TBRQ_TPD(he_dev->tbrq_head)) {
1829 TBRQ_TPD(he_dev->tbrq_head));
1833 if (TBRQ_EOS(he_dev->tbrq_head)) {
1835 he_mkcid(he_dev, tpd->vcc->vpi, tpd->vcc->vci));
1844 dma_unmap_single(&he_dev->pci_dev->dev,
1853 if (tpd->skb) { /* && !TBRQ_MULTIPLE(he_dev->tbrq_head) */
1862 dma_pool_free(he_dev->tpd_pool, tpd, TPD_ADDR(tpd->status));
1863 he_dev->tbrq_head = (struct he_tbrq *)
1864 ((unsigned long) he_dev->tbrq_base |
1865 TBRQ_MASK(he_dev->tbrq_head + 1));
1869 if (updated > he_dev->tbrq_peak)
1870 he_dev->tbrq_peak = updated;
1872 he_writel(he_dev, TBRQ_MASK(he_dev->tbrq_head),
1878 he_service_rbpl(struct he_dev *he_dev, int group)
1887 rbpl_head = (struct he_rbp *) ((unsigned long)he_dev->rbpl_base |
1888 RBPL_MASK(he_readl(he_dev, G0_RBPL_S)));
1891 new_tail = (struct he_rbp *) ((unsigned long)he_dev->rbpl_base |
1892 RBPL_MASK(he_dev->rbpl_tail+1));
1898 i = find_next_zero_bit(he_dev->rbpl_table, RBPL_TABLE_SIZE, he_dev->rbpl_hint);
1900 i = find_first_zero_bit(he_dev->rbpl_table, RBPL_TABLE_SIZE);
1904 he_dev->rbpl_hint = i + 1;
1906 heb = dma_pool_alloc(he_dev->rbpl_pool, GFP_ATOMIC, &mapping);
1910 list_add(&heb->entry, &he_dev->rbpl_outstanding);
1911 he_dev->rbpl_virt[i] = heb;
1912 set_bit(i, he_dev->rbpl_table);
1916 he_dev->rbpl_tail = new_tail;
1921 he_writel(he_dev, RBPL_MASK(he_dev->rbpl_tail), G0_RBPL_T);
1928 struct he_dev *he_dev = (struct he_dev *) data;
1933 spin_lock_irqsave(&he_dev->global_lock, flags);
1935 while (he_dev->irq_head != he_dev->irq_tail) {
1938 type = ITYPE_TYPE(he_dev->irq_head->isw);
1939 group = ITYPE_GROUP(he_dev->irq_head->isw);
1946 if (he_service_rbrq(he_dev, group))
1947 he_service_rbpl(he_dev, group);
1953 he_service_tbrq(he_dev, group);
1956 he_service_rbpl(he_dev, group);
1964 spin_unlock_irqrestore(&he_dev->global_lock, flags);
1965 if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->interrupt)
1966 he_dev->atm_dev->phy->interrupt(he_dev->atm_dev);
1967 spin_lock_irqsave(&he_dev->global_lock, flags);
1976 hprintk("abort 0x%x\n", he_readl(he_dev, ABORT_ADDR));
1983 HPRINTK("isw not updated 0x%x\n", he_dev->irq_head->isw);
1985 he_service_rbrq(he_dev, 0);
1986 he_service_rbpl(he_dev, 0);
1987 he_service_tbrq(he_dev, 0);
1990 hprintk("bad isw 0x%x?\n", he_dev->irq_head->isw);
1993 he_dev->irq_head->isw = ITYPE_INVALID;
1995 he_dev->irq_head = (struct he_irq *) NEXT_ENTRY(he_dev->irq_base, he_dev->irq_head, IRQ_MASK);
1999 if (updated > he_dev->irq_peak)
2000 he_dev->irq_peak = updated;
2002 he_writel(he_dev,
2005 IRQ_TAIL(he_dev->irq_tail), IRQ0_HEAD);
2006 (void) he_readl(he_dev, INT_FIFO); /* 8.1.2 controller errata; flush posted writes */
2008 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2015 struct he_dev *he_dev = (struct he_dev * )dev_id;
2018 if (he_dev == NULL)
2021 spin_lock_irqsave(&he_dev->global_lock, flags);
2023 he_dev->irq_tail = (struct he_irq *) (((unsigned long)he_dev->irq_base) |
2024 (*he_dev->irq_tailoffset << 2));
2026 if (he_dev->irq_tail == he_dev->irq_head) {
2028 he_dev->irq_tail = (struct he_irq *) ((unsigned long)he_dev->irq_base |
2029 ((he_readl(he_dev, IRQ0_BASE) & IRQ_MASK) << 2));
2030 (void) he_readl(he_dev, INT_FIFO); /* 8.1.2 controller errata */
2034 if (he_dev->irq_head == he_dev->irq_tail /* && !IRQ_PENDING */)
2038 if (he_dev->irq_head != he_dev->irq_tail) {
2040 tasklet_schedule(&he_dev->tasklet);
2041 he_writel(he_dev, INT_CLEAR_A, INT_FIFO); /* clear interrupt */
2042 (void) he_readl(he_dev, INT_FIFO); /* flush posted writes */
2044 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2050 __enqueue_tpd(struct he_dev *he_dev, struct he_tpd *tpd, unsigned cid)
2055 tpd, cid, he_dev->tpdrq_tail);
2057 /* new_tail = he_dev->tpdrq_tail; */
2058 new_tail = (struct he_tpdrq *) ((unsigned long) he_dev->tpdrq_base |
2059 TPDRQ_MASK(he_dev->tpdrq_tail+1));
2068 if (new_tail == he_dev->tpdrq_head) {
2069 he_dev->tpdrq_head = (struct he_tpdrq *)
2070 (((unsigned long)he_dev->tpdrq_base) |
2071 TPDRQ_MASK(he_readl(he_dev, TPDRQ_B_H)));
2073 if (new_tail == he_dev->tpdrq_head) {
2085 dma_unmap_single(&he_dev->pci_dev->dev,
2097 dma_pool_free(he_dev->tpd_pool, tpd, TPD_ADDR(tpd->status));
2103 list_add_tail(&tpd->entry, &he_dev->outstanding_tpds);
2104 he_dev->tpdrq_tail->tpd = TPD_ADDR(tpd->status);
2105 he_dev->tpdrq_tail->cid = cid;
2108 he_dev->tpdrq_tail = new_tail;
2110 he_writel(he_dev, TPDRQ_MASK(he_dev->tpdrq_tail), TPDRQ_T);
2111 (void) he_readl(he_dev, TPDRQ_T); /* flush posted writes */
2118 struct he_dev *he_dev = HE_DEV(vcc->dev);
2132 cid = he_mkcid(he_dev, vpi, vci);
2154 pcr_goal = he_dev->atm_dev->link_rate;
2174 spin_lock_irqsave(&he_dev->global_lock, flags);
2175 tsr0 = he_readl_tsr0(he_dev, cid);
2176 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2196 if ((he_dev->total_bw + pcr_goal)
2197 > (he_dev->atm_dev->link_rate * 9 / 10))
2203 spin_lock_irqsave(&he_dev->global_lock, flags); /* also protects he_dev->cs_stper[] */
2207 if (he_dev->cs_stper[reg].inuse == 0 ||
2208 he_dev->cs_stper[reg].pcr == pcr_goal)
2213 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2217 he_dev->total_bw += pcr_goal;
2220 ++he_dev->cs_stper[reg].inuse;
2221 he_dev->cs_stper[reg].pcr = pcr_goal;
2223 clock = he_is622(he_dev) ? 66667000 : 50000000;
2229 he_writel_mbox(he_dev, rate_to_atmf(period/2),
2231 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2242 spin_lock_irqsave(&he_dev->global_lock, flags);
2244 he_writel_tsr0(he_dev, tsr0, cid);
2245 he_writel_tsr4(he_dev, tsr4 | 1, cid);
2246 he_writel_tsr1(he_dev, TSR1_MCR(rate_to_atmf(0)) |
2248 he_writel_tsr2(he_dev, TSR2_ACR(rate_to_atmf(pcr_goal)), cid);
2249 he_writel_tsr9(he_dev, TSR9_OPEN_CONN, cid);
2251 he_writel_tsr3(he_dev, 0x0, cid);
2252 he_writel_tsr5(he_dev, 0x0, cid);
2253 he_writel_tsr6(he_dev, 0x0, cid);
2254 he_writel_tsr7(he_dev, 0x0, cid);
2255 he_writel_tsr8(he_dev, 0x0, cid);
2256 he_writel_tsr10(he_dev, 0x0, cid);
2257 he_writel_tsr11(he_dev, 0x0, cid);
2258 he_writel_tsr12(he_dev, 0x0, cid);
2259 he_writel_tsr13(he_dev, 0x0, cid);
2260 he_writel_tsr14(he_dev, 0x0, cid);
2261 (void) he_readl_tsr0(he_dev, cid); /* flush posted writes */
2262 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2283 spin_lock_irqsave(&he_dev->global_lock, flags);
2285 rsr0 = he_readl_rsr0(he_dev, cid);
2287 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2304 he_writel_rsr4(he_dev, rsr4, cid);
2305 he_writel_rsr1(he_dev, rsr1, cid);
2308 he_writel_rsr0(he_dev,
2310 (void) he_readl_rsr0(he_dev, cid); /* flush posted writes */
2312 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2332 struct he_dev *he_dev = HE_DEV(vcc->dev);
2342 cid = he_mkcid(he_dev, vcc->vpi, vcc->vci);
2353 spin_lock_irqsave(&he_dev->global_lock, flags);
2354 while (he_readl(he_dev, RCC_STAT) & RCC_BUSY) {
2362 he_writel_rsr0(he_dev, RSR0_CLOSE_CONN, cid);
2363 (void) he_readl_rsr0(he_dev, cid); /* flush posted writes */
2364 he_writel_mbox(he_dev, cid, RXCON_CLOSE);
2365 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2408 spin_lock_irqsave(&he_dev->global_lock, flags);
2409 he_writel_tsr4_upper(he_dev, TSR4_FLUSH_CONN, cid);
2414 he_writel_tsr1(he_dev,
2419 he_writel_tsr14_upper(he_dev, TSR14_DELETE, cid);
2422 (void) he_readl_tsr4(he_dev, cid); /* flush posted writes */
2424 tpd = __alloc_tpd(he_dev);
2436 __enqueue_tpd(he_dev, tpd, cid);
2437 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2444 spin_lock_irqsave(&he_dev->global_lock, flags);
2451 while (!((tsr4 = he_readl_tsr4(he_dev, cid)) & TSR4_SESSION_ENDED)) {
2456 while (TSR0_CONN_STATE(tsr0 = he_readl_tsr0(he_dev, cid)) != 0) {
2468 if (he_dev->cs_stper[reg].inuse == 0)
2471 --he_dev->cs_stper[reg].inuse;
2473 he_dev->total_bw -= he_dev->cs_stper[reg].pcr;
2475 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2489 struct he_dev *he_dev = HE_DEV(vcc->dev);
2490 unsigned cid = he_mkcid(he_dev, vcc->vpi, vcc->vci);
2522 spin_lock_irqsave(&he_dev->global_lock, flags);
2524 tpd = __alloc_tpd(he_dev);
2531 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2551 tpd->iovec[slot].addr = dma_map_single(&he_dev->pci_dev->dev, skb->data,
2565 __enqueue_tpd(he_dev, tpd, cid);
2566 tpd = __alloc_tpd(he_dev);
2573 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2580 tpd->iovec[slot].addr = skb_frag_dma_map(&he_dev->pci_dev->dev,
2589 tpd->address0 = dma_map_single(&he_dev->pci_dev->dev, skb->data, skb->len, DMA_TO_DEVICE);
2599 __enqueue_tpd(he_dev, tpd, cid);
2600 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2611 struct he_dev *he_dev = HE_DEV(atm_dev);
2624 spin_lock_irqsave(&he_dev->global_lock, flags);
2632 reg.val = he_readl(he_dev, reg.addr);
2636 he_readl_rcm(he_dev, reg.addr);
2640 he_readl_tcm(he_dev, reg.addr);
2644 he_readl_mbox(he_dev, reg.addr);
2650 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2673 struct he_dev *he_dev = HE_DEV(atm_dev);
2677 spin_lock_irqsave(&he_dev->global_lock, flags);
2678 he_writel(he_dev, val, FRAMER + (addr*4));
2679 (void) he_readl(he_dev, FRAMER + (addr*4)); /* flush posted writes */
2680 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2688 struct he_dev *he_dev = HE_DEV(atm_dev);
2691 spin_lock_irqsave(&he_dev->global_lock, flags);
2692 reg = he_readl(he_dev, FRAMER + (addr*4));
2693 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2703 struct he_dev *he_dev = HE_DEV(dev);
2719 he_dev->prod_id, he_dev->media & 0x40 ? "SM" : "MM");
2724 spin_lock_irqsave(&he_dev->global_lock, flags);
2725 mcc += he_readl(he_dev, MCC);
2726 oec += he_readl(he_dev, OEC);
2727 dcc += he_readl(he_dev, DCC);
2728 cec += he_readl(he_dev, CEC);
2729 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2737 CONFIG_IRQ_SIZE, he_dev->irq_peak);
2745 CONFIG_RBRQ_SIZE, he_dev->rbrq_peak);
2749 CONFIG_TBRQ_SIZE, he_dev->tbrq_peak);
2753 rbpl_head = RBPL_MASK(he_readl(he_dev, G0_RBPL_S));
2754 rbpl_tail = RBPL_MASK(he_readl(he_dev, G0_RBPL_T));
2772 he_dev->cs_stper[i].pcr,
2773 he_dev->cs_stper[i].inuse);
2777 he_dev->total_bw, he_dev->atm_dev->link_rate * 10 / 9);
2784 static u8 read_prom_byte(struct he_dev *he_dev, int addr)
2790 val = readl(he_dev->membase + HOST_CNTL);
2795 he_writel(he_dev, val, HOST_CNTL);
2799 he_writel(he_dev, val | readtab[i], HOST_CNTL);
2805 he_writel(he_dev, val | clocktab[j++] | (((addr >> i) & 1) << 9), HOST_CNTL);
2807 he_writel(he_dev, val | clocktab[j++] | (((addr >> i) & 1) << 9), HOST_CNTL);
2814 he_writel(he_dev, val, HOST_CNTL);
2818 he_writel(he_dev, val | clocktab[j++], HOST_CNTL);
2820 tmp_read = he_readl(he_dev, HOST_CNTL);
2823 he_writel(he_dev, val | clocktab[j++], HOST_CNTL);
2827 he_writel(he_dev, val | ID_CS, HOST_CNTL);