Lines Matching defs:pdev
1107 struct pci_dev *pdev = to_pci_dev(nvmeq->dev->dev);
1111 disable_irq(pci_irq_vector(pdev, nvmeq->cq_vector));
1113 enable_irq(pci_irq_vector(pdev, nvmeq->cq_vector));
1496 struct pci_dev *pdev = to_pci_dev(dev->dev);
1499 nvmeq->sq_cmds = pci_alloc_p2pmem(pdev, SQ_SIZE(nvmeq));
1501 nvmeq->sq_dma_addr = pci_p2pmem_virt_to_bus(pdev,
1508 pci_free_p2pmem(pdev, nvmeq->sq_cmds, SQ_SIZE(nvmeq));
1556 struct pci_dev *pdev = to_pci_dev(nvmeq->dev->dev);
1560 return pci_request_irq(pdev, nvmeq->cq_vector, nvme_irq_check,
1563 return pci_request_irq(pdev, nvmeq->cq_vector, nvme_irq,
1697 struct pci_dev *pdev = to_pci_dev(dev->dev);
1701 if (size > pci_resource_len(pdev, 0))
1705 dev->bar = ioremap(pci_resource_start(pdev, 0), size);
1827 struct pci_dev *pdev = to_pci_dev(dev->dev);
1844 bar_size = pci_resource_len(pdev, bar);
1855 (pci_bus_address(pdev, bar) + offset),
1867 if (pci_p2pdma_add_resource(pdev, bar, size, offset)) {
1878 pci_p2pmem_publish(pdev, true);
2212 struct pci_dev *pdev = to_pci_dev(dev->dev);
2242 return pci_alloc_irq_vectors_affinity(pdev, 1, irq_queues,
2260 struct pci_dev *pdev = to_pci_dev(dev->dev);
2291 pci_free_irq(pdev, 0, adminq);
2319 pci_free_irq(pdev, 0, adminq);
2325 pci_free_irq_vectors(pdev);
2471 struct pci_dev *pdev = to_pci_dev(dev->dev);
2473 if (pci_enable_device_mem(pdev))
2476 pci_set_master(pdev);
2488 result = pci_alloc_irq_vectors(pdev, 1, 1, PCI_IRQ_ALL_TYPES);
2513 if (pdev->vendor == PCI_VENDOR_ID_APPLE && pdev->device == 0x2001) {
2518 } else if (pdev->vendor == PCI_VENDOR_ID_SAMSUNG &&
2519 (pdev->device == 0xa821 || pdev->device == 0xa822) &&
2540 pci_save_state(pdev);
2548 pci_free_irq_vectors(pdev);
2550 pci_disable_device(pdev);
2563 struct pci_dev *pdev = to_pci_dev(dev->dev);
2566 if (!pci_is_enabled(pdev) || !pci_device_is_present(pdev))
2568 if (pdev->error_state != pci_channel_io_normal)
2578 struct pci_dev *pdev = to_pci_dev(dev->dev);
2584 if (pci_is_enabled(pdev))
2603 pci_free_irq_vectors(pdev);
2604 if (pci_is_enabled(pdev))
2605 pci_disable_device(pdev);
2811 struct pci_dev *pdev = to_pci_dev(to_nvme_dev(ctrl)->dev);
2813 return snprintf(buf, size, "%s\n", dev_name(&pdev->dev));
2818 struct pci_dev *pdev = to_pci_dev(to_nvme_dev(ctrl)->dev);
2823 pdev->vendor, pdev->device,
2854 struct pci_dev *pdev = to_pci_dev(dev->dev);
2856 if (pci_request_mem_regions(pdev, "nvme"))
2864 pci_release_mem_regions(pdev);
2868 static unsigned long check_vendor_combination_bug(struct pci_dev *pdev)
2870 if (pdev->vendor == 0x144d && pdev->device == 0xa802) {
2883 } else if (pdev->vendor == 0x144d && pdev->device == 0xa804) {
2894 } else if ((pdev->vendor == 0x144d && (pdev->device == 0xa801 ||
2895 pdev->device == 0xa808 || pdev->device == 0xa809)) ||
2896 (pdev->vendor == 0x1e0f && pdev->device == 0x0001)) {
2906 } else if (pdev->vendor == 0x2646 && (pdev->device == 0x2263 ||
2907 pdev->device == 0x500f)) {
2923 static struct nvme_dev *nvme_pci_alloc_dev(struct pci_dev *pdev,
2927 int node = dev_to_node(&pdev->dev);
2945 dev->dev = get_device(&pdev->dev);
2947 quirks |= check_vendor_combination_bug(pdev);
2950 acpi_storage_d3(&pdev->dev)) {
2955 dev_info(&pdev->dev,
2959 ret = nvme_init_ctrl(&dev->ctrl, &pdev->dev, &nvme_pci_ctrl_ops,
2965 dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(48));
2967 dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
2968 dma_set_min_align_mask(&pdev->dev, NVME_CTRL_PAGE_SIZE - 1);
2969 dma_set_max_seg_size(&pdev->dev, 0xffffffff);
2976 NVME_MAX_KB_SZ << 1, dma_opt_mapping_size(&pdev->dev) >> 9);
2994 static int nvme_probe(struct pci_dev *pdev, const struct pci_device_id *id)
2999 dev = nvme_pci_alloc_dev(pdev, id);
3015 dev_info(dev->ctrl.device, "pci function %s\n", dev_name(&pdev->dev));
3067 pci_set_drvdata(pdev, dev);
3093 static void nvme_reset_prepare(struct pci_dev *pdev)
3095 struct nvme_dev *dev = pci_get_drvdata(pdev);
3106 static void nvme_reset_done(struct pci_dev *pdev)
3108 struct nvme_dev *dev = pci_get_drvdata(pdev);
3114 static void nvme_shutdown(struct pci_dev *pdev)
3116 struct nvme_dev *dev = pci_get_drvdata(pdev);
3126 static void nvme_remove(struct pci_dev *pdev)
3128 struct nvme_dev *dev = pci_get_drvdata(pdev);
3131 pci_set_drvdata(pdev, NULL);
3133 if (!pci_device_is_present(pdev)) {
3181 struct pci_dev *pdev = to_pci_dev(dev);
3182 struct nvme_dev *ndev = pci_get_drvdata(pdev);
3202 !pcie_aspm_enabled(pdev) ||
3233 pci_save_state(pdev);
3241 pci_load_saved_state(pdev, NULL);
3264 struct pci_dev *pdev = to_pci_dev(dev);
3265 struct nvme_dev *ndev = pci_get_drvdata(pdev);
3280 static pci_ers_result_t nvme_error_detected(struct pci_dev *pdev,
3283 struct nvme_dev *dev = pci_get_drvdata(pdev);
3310 static pci_ers_result_t nvme_slot_reset(struct pci_dev *pdev)
3312 struct nvme_dev *dev = pci_get_drvdata(pdev);
3315 pci_restore_state(pdev);
3321 static void nvme_error_resume(struct pci_dev *pdev)
3323 struct nvme_dev *dev = pci_get_drvdata(pdev);