Lines Matching refs:rtlpci
65 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
73 ppsc->const_amdpci_aspm = rtlpci->const_amdpci_aspm;
74 switch (rtlpci->const_pci_aspm) {
112 switch (rtlpci->const_hwsw_rfoff_d3) {
130 switch (rtlpci->const_support_pciaspm) {
147 rtlpci->const_support_pciaspm);
154 pci_read_config_byte(rtlpci->pdev, 0x80, &init_aspm);
164 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
172 pcie_capability_clear_and_set_word(rtlpci->pdev, PCI_EXP_LNKCTL,
182 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
187 pcie_capability_clear_and_set_word(rtlpci->pdev, PCI_EXP_LNKCTL,
201 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
224 pci_read_config_byte(rtlpci->pdev, 0x80, &tmp_u1b);
243 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
258 aspmlevel = rtlpci->const_devicepci_aspm_setting;
279 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
285 pci_write_config_byte(rtlpci->pdev, 0xe0, 0xa0);
287 pci_read_config_byte(rtlpci->pdev, 0xe0, &offset_e0);
290 pci_read_config_dword(rtlpci->pdev, 0xe4, &offset_e4);
437 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
449 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue];
484 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
486 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[prio];
505 dma_unmap_single(&rtlpci->pdev->dev,
587 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
596 skb = dev_alloc_skb(rtlpci->rxbuffersize);
603 dma_map_single(&rtlpci->pdev->dev, skb_tail_pointer(skb),
604 rtlpci->rxbuffersize, DMA_FROM_DEVICE);
606 if (dma_mapping_error(&rtlpci->pdev->dev, bufferaddress))
608 rtlpci->rx_ring[rxring_idx].rx_buf[desc_idx] = skb;
620 (u8 *)&rtlpci->rxbuffersize);
667 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
671 rtlpci->sys_irq_mask);
677 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
680 unsigned int count = rtlpci->rxringcount;
701 struct sk_buff *skb = rtlpci->rx_ring[rxring_idx].rx_buf[
702 rtlpci->rx_ring[rxring_idx].idx];
712 buffer_desc = &rtlpci->rx_ring[rxring_idx].buffer_desc[
713 rtlpci->rx_ring[rxring_idx].idx];
716 pdesc = &rtlpci->rx_ring[rxring_idx].desc[
717 rtlpci->rx_ring[rxring_idx].idx];
730 dma_unmap_single(&rtlpci->pdev->dev, *((dma_addr_t *)skb->cb),
731 rtlpci->rxbuffersize, DMA_FROM_DEVICE);
734 new_skb = dev_alloc_skb(rtlpci->rxbuffersize);
821 rtlpci->rx_ring[hw_queue].next_rx_rp += 1;
822 rtlpci->rx_ring[hw_queue].next_rx_rp %=
827 rtlpci->rx_ring[hw_queue].next_rx_rp);
838 rtlpci->rx_ring[rxring_idx].idx);
842 rtlpci->rx_ring[rxring_idx].idx);
843 if (rtlpci->rx_ring[rxring_idx].idx ==
844 rtlpci->rxringcount - 1)
850 rtlpci->rx_ring[rxring_idx].idx =
851 (rtlpci->rx_ring[rxring_idx].idx + 1) %
852 rtlpci->rxringcount;
859 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
867 if (rtlpci->irq_enabled == 0)
1030 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1044 ring = &rtlpci->tx_ring[BEACON_QUEUE];
1051 dma_unmap_single(&rtlpci->pdev->dev,
1086 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1100 rtlpci->txringcount[i] = desc_num;
1105 rtlpci->txringcount[BEACON_QUEUE] = 2;
1112 rtlpci->txringcount[BE_QUEUE] = RT_TXDESC_NUM_BE_QUEUE;
1114 rtlpci->rxbuffersize = 9100; /*2048/1024; */
1115 rtlpci->rxringcount = RTL_PCI_MAX_RX_COUNT; /*64; */
1123 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1126 rtlpci->up_first_time = true;
1127 rtlpci->being_init_adapter = false;
1130 rtlpci->pdev = pdev;
1150 rtlpci->acm_method = EACMWAY2_SW;
1163 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1174 dma_alloc_coherent(&rtlpci->pdev->dev,
1184 rtlpci->tx_ring[prio].buffer_desc = buffer_desc;
1185 rtlpci->tx_ring[prio].buffer_desc_dma = buffer_desc_dma;
1187 rtlpci->tx_ring[prio].cur_tx_rp = 0;
1188 rtlpci->tx_ring[prio].cur_tx_wp = 0;
1192 desc = dma_alloc_coherent(&rtlpci->pdev->dev, sizeof(*desc) * entries,
1200 rtlpci->tx_ring[prio].desc = desc;
1201 rtlpci->tx_ring[prio].dma = desc_dma;
1203 rtlpci->tx_ring[prio].idx = 0;
1204 rtlpci->tx_ring[prio].entries = entries;
1205 skb_queue_head_init(&rtlpci->tx_ring[prio].queue);
1228 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1235 rtlpci->rx_ring[rxring_idx].buffer_desc =
1236 dma_alloc_coherent(&rtlpci->pdev->dev,
1237 sizeof(*rtlpci->rx_ring[rxring_idx].buffer_desc) *
1238 rtlpci->rxringcount,
1239 &rtlpci->rx_ring[rxring_idx].dma, GFP_KERNEL);
1240 if (!rtlpci->rx_ring[rxring_idx].buffer_desc ||
1241 (ulong)rtlpci->rx_ring[rxring_idx].buffer_desc & 0xFF) {
1247 rtlpci->rx_ring[rxring_idx].idx = 0;
1248 for (i = 0; i < rtlpci->rxringcount; i++) {
1249 entry = &rtlpci->rx_ring[rxring_idx].buffer_desc[i];
1258 rtlpci->rx_ring[rxring_idx].desc =
1259 dma_alloc_coherent(&rtlpci->pdev->dev,
1260 sizeof(*rtlpci->rx_ring[rxring_idx].desc) *
1261 rtlpci->rxringcount,
1262 &rtlpci->rx_ring[rxring_idx].dma, GFP_KERNEL);
1263 if (!rtlpci->rx_ring[rxring_idx].desc ||
1264 (unsigned long)rtlpci->rx_ring[rxring_idx].desc & 0xFF) {
1270 rtlpci->rx_ring[rxring_idx].idx = 0;
1272 for (i = 0; i < rtlpci->rxringcount; i++) {
1273 entry = &rtlpci->rx_ring[rxring_idx].desc[i];
1289 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1290 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[prio];
1302 dma_unmap_single(&rtlpci->pdev->dev,
1311 dma_free_coherent(&rtlpci->pdev->dev,
1316 dma_free_coherent(&rtlpci->pdev->dev,
1326 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1330 for (i = 0; i < rtlpci->rxringcount; i++) {
1331 struct sk_buff *skb = rtlpci->rx_ring[rxring_idx].rx_buf[i];
1335 dma_unmap_single(&rtlpci->pdev->dev, *((dma_addr_t *)skb->cb),
1336 rtlpci->rxbuffersize, DMA_FROM_DEVICE);
1342 dma_free_coherent(&rtlpci->pdev->dev,
1343 sizeof(*rtlpci->rx_ring[rxring_idx].buffer_desc) *
1344 rtlpci->rxringcount,
1345 rtlpci->rx_ring[rxring_idx].buffer_desc,
1346 rtlpci->rx_ring[rxring_idx].dma);
1347 rtlpci->rx_ring[rxring_idx].buffer_desc = NULL;
1349 dma_free_coherent(&rtlpci->pdev->dev,
1350 sizeof(*rtlpci->rx_ring[rxring_idx].desc) *
1351 rtlpci->rxringcount,
1352 rtlpci->rx_ring[rxring_idx].desc,
1353 rtlpci->rx_ring[rxring_idx].dma);
1354 rtlpci->rx_ring[rxring_idx].desc = NULL;
1360 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1374 ret = _rtl_pci_init_tx_ring(hw, i, rtlpci->txringcount[i]);
1386 if (rtlpci->tx_ring[i].desc ||
1387 rtlpci->tx_ring[i].buffer_desc)
1411 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1424 rtlpci->rx_ring[rxring_idx].desc) {
1427 rtlpci->rx_ring[rxring_idx].idx = 0;
1428 for (i = 0; i < rtlpci->rxringcount; i++) {
1429 entry = &rtlpci->rx_ring[rxring_idx].desc[i];
1434 sizeof(*rtlpci->rx_ring
1449 (u8 *)&rtlpci->rxbuffersize);
1459 rtlpci->rx_ring[rxring_idx].idx = 0;
1467 if (rtlpci->tx_ring[i].desc ||
1468 rtlpci->tx_ring[i].buffer_desc) {
1469 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[i];
1481 dma_unmap_single(&rtlpci->pdev->dev,
1490 rtlpci->tx_ring[i].cur_tx_rp = 0;
1491 rtlpci->tx_ring[i].cur_tx_wp = 0;
1495 ring->entries = rtlpci->txringcount[i];
1558 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1581 ring = &rtlpci->tx_ring[hw_queue];
1693 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1697 synchronize_irq(rtlpci->pdev->irq);
1723 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1732 rtlpci->driver_is_goingto_unload = false;
1763 rtlpci->up_first_time = false;
1772 struct rtl_pci *rtlpci = rtl_pcidev(rtl_pcipriv(hw));
1789 rtlpci->driver_is_goingto_unload = true;
2025 struct rtl_pci *rtlpci = rtl_pcidev(pcipriv);
2028 ret = pci_enable_msi(rtlpci->pdev);
2032 ret = request_irq(rtlpci->pdev->irq, &_rtl_pci_interrupt,
2035 pci_disable_msi(rtlpci->pdev);
2039 rtlpci->using_msi = true;
2050 struct rtl_pci *rtlpci = rtl_pcidev(pcipriv);
2053 ret = request_irq(rtlpci->pdev->irq, &_rtl_pci_interrupt,
2058 rtlpci->using_msi = false;
2067 struct rtl_pci *rtlpci = rtl_pcidev(pcipriv);
2070 if (rtlpci->msi_support) {
2102 struct rtl_pci *rtlpci;
2249 rtlpci = rtl_pcidev(pcipriv);
2257 rtlpci->irq_alloc = 1;
2287 struct rtl_pci *rtlpci = rtl_pcidev(pcipriv);
2314 if (rtlpci->irq_alloc) {
2315 free_irq(rtlpci->pdev->irq, hw);
2316 rtlpci->irq_alloc = 0;
2319 if (rtlpci->using_msi)
2320 pci_disable_msi(rtlpci->pdev);