Lines Matching refs:vmd
114 static inline unsigned int index_from_irqs(struct vmd_dev *vmd,
117 return irqs - vmd->irqs;
125 * handlers sharing that VMD IRQ. The vmd irq_domain provides the operations
132 struct vmd_dev *vmd = irq_data_get_irq_handler_data(data);
136 MSI_ADDR_DEST_ID(index_from_irqs(vmd, irq));
200 static struct vmd_irq_list *vmd_next_irq(struct vmd_dev *vmd, struct msi_desc *desc)
205 if (vmd->msix_count == 1)
206 return &vmd->irqs[0];
216 return &vmd->irqs[0];
220 for (i = 1; i < vmd->msix_count; i++)
221 if (vmd->irqs[i].count < vmd->irqs[best].count)
223 vmd->irqs[best].count++;
226 return &vmd->irqs[best];
234 struct vmd_dev *vmd = vmd_from_bus(msi_desc_to_pci_dev(desc)->bus);
242 vmdirq->irq = vmd_next_irq(vmd, desc);
244 index = index_from_irqs(vmd, vmdirq->irq);
245 vector = pci_irq_vector(vmd->dev, index);
248 handle_untracked_irq, vmd, NULL);
272 struct vmd_dev *vmd = vmd_from_bus(pdev->bus);
274 if (nvec > vmd->msix_count)
275 return vmd->msix_count;
301 static int vmd_create_irq_domain(struct vmd_dev *vmd)
305 fn = irq_domain_alloc_named_id_fwnode("VMD-MSI", vmd->sysdata.domain);
309 vmd->irq_domain = pci_msi_create_irq_domain(fn, &vmd_msi_domain_info, NULL);
310 if (!vmd->irq_domain) {
318 static void vmd_remove_irq_domain(struct vmd_dev *vmd)
320 if (vmd->irq_domain) {
321 struct fwnode_handle *fn = vmd->irq_domain->fwnode;
323 irq_domain_remove(vmd->irq_domain);
328 static char __iomem *vmd_cfg_addr(struct vmd_dev *vmd, struct pci_bus *bus,
331 char __iomem *addr = vmd->cfgbar +
332 ((bus->number - vmd->busn_start) << 20) +
335 if ((addr - vmd->cfgbar) + len >=
336 resource_size(&vmd->dev->resource[VMD_CFGBAR]))
349 struct vmd_dev *vmd = vmd_from_bus(bus);
350 char __iomem *addr = vmd_cfg_addr(vmd, bus, devfn, reg, len);
357 spin_lock_irqsave(&vmd->cfg_lock, flags);
372 spin_unlock_irqrestore(&vmd->cfg_lock, flags);
384 struct vmd_dev *vmd = vmd_from_bus(bus);
385 char __iomem *addr = vmd_cfg_addr(vmd, bus, devfn, reg, len);
392 spin_lock_irqsave(&vmd->cfg_lock, flags);
410 spin_unlock_irqrestore(&vmd->cfg_lock, flags);
419 static void vmd_attach_resources(struct vmd_dev *vmd)
421 vmd->dev->resource[VMD_MEMBAR1].child = &vmd->resources[1];
422 vmd->dev->resource[VMD_MEMBAR2].child = &vmd->resources[2];
425 static void vmd_detach_resources(struct vmd_dev *vmd)
427 vmd->dev->resource[VMD_MEMBAR1].child = NULL;
428 vmd->dev->resource[VMD_MEMBAR2].child = NULL;
447 static int vmd_get_phys_offsets(struct vmd_dev *vmd, bool native_hint,
451 struct pci_dev *dev = vmd->dev;
501 static int vmd_get_bus_number_start(struct vmd_dev *vmd)
503 struct pci_dev *dev = vmd->dev;
512 vmd->busn_start = 0;
515 vmd->busn_start = 128;
518 vmd->busn_start = 224;
544 static int vmd_alloc_irqs(struct vmd_dev *vmd)
546 struct pci_dev *dev = vmd->dev;
549 vmd->msix_count = pci_msix_vec_count(dev);
550 if (vmd->msix_count < 0)
553 vmd->msix_count = pci_alloc_irq_vectors(dev, 1, vmd->msix_count,
555 if (vmd->msix_count < 0)
556 return vmd->msix_count;
558 vmd->irqs = devm_kcalloc(&dev->dev, vmd->msix_count, sizeof(*vmd->irqs),
560 if (!vmd->irqs)
563 for (i = 0; i < vmd->msix_count; i++) {
564 err = init_srcu_struct(&vmd->irqs[i].srcu);
568 INIT_LIST_HEAD(&vmd->irqs[i].irq_list);
571 "vmd", &vmd->irqs[i]);
579 static int vmd_enable_domain(struct vmd_dev *vmd, unsigned long features)
581 struct pci_sysdata *sd = &vmd->sysdata;
599 ret = vmd_get_phys_offsets(vmd, true, &offset[0], &offset[1]);
603 ret = vmd_get_phys_offsets(vmd, false, &offset[0], &offset[1]);
613 ret = vmd_get_bus_number_start(vmd);
618 res = &vmd->dev->resource[VMD_CFGBAR];
619 vmd->resources[0] = (struct resource) {
621 .start = vmd->busn_start,
622 .end = vmd->busn_start + (resource_size(res) >> 20) - 1,
643 res = &vmd->dev->resource[VMD_MEMBAR1];
648 vmd->resources[1] = (struct resource) {
656 res = &vmd->dev->resource[VMD_MEMBAR2];
661 vmd->resources[2] = (struct resource) {
669 sd->vmd_dev = vmd->dev;
674 sd->node = pcibus_to_node(vmd->dev->bus);
676 ret = vmd_create_irq_domain(vmd);
684 irq_domain_update_bus_token(vmd->irq_domain, DOMAIN_BUS_VMD_MSI);
686 pci_add_resource(&resources, &vmd->resources[0]);
687 pci_add_resource_offset(&resources, &vmd->resources[1], offset[0]);
688 pci_add_resource_offset(&resources, &vmd->resources[2], offset[1]);
690 vmd->bus = pci_create_root_bus(&vmd->dev->dev, vmd->busn_start,
692 if (!vmd->bus) {
694 vmd_remove_irq_domain(vmd);
698 vmd_attach_resources(vmd);
699 if (vmd->irq_domain)
700 dev_set_msi_domain(&vmd->bus->dev, vmd->irq_domain);
702 pci_scan_child_bus(vmd->bus);
703 pci_assign_unassigned_bus_resources(vmd->bus);
710 list_for_each_entry(child, &vmd->bus->children, node)
713 pci_bus_add_devices(vmd->bus);
715 WARN(sysfs_create_link(&vmd->dev->dev.kobj, &vmd->bus->dev.kobj,
722 struct vmd_dev *vmd;
728 vmd = devm_kzalloc(&dev->dev, sizeof(*vmd), GFP_KERNEL);
729 if (!vmd)
732 vmd->dev = dev;
737 vmd->cfgbar = pcim_iomap(dev, VMD_CFGBAR, 0);
738 if (!vmd->cfgbar)
746 err = vmd_alloc_irqs(vmd);
750 spin_lock_init(&vmd->cfg_lock);
751 pci_set_drvdata(dev, vmd);
752 err = vmd_enable_domain(vmd, (unsigned long) id->driver_data);
756 dev_info(&vmd->dev->dev, "Bound to PCI domain %04x\n",
757 vmd->sysdata.domain);
761 static void vmd_cleanup_srcu(struct vmd_dev *vmd)
765 for (i = 0; i < vmd->msix_count; i++)
766 cleanup_srcu_struct(&vmd->irqs[i].srcu);
771 struct vmd_dev *vmd = pci_get_drvdata(dev);
773 sysfs_remove_link(&vmd->dev->dev.kobj, "domain");
774 pci_stop_root_bus(vmd->bus);
775 pci_remove_root_bus(vmd->bus);
776 vmd_cleanup_srcu(vmd);
777 vmd_detach_resources(vmd);
778 vmd_remove_irq_domain(vmd);
785 struct vmd_dev *vmd = pci_get_drvdata(pdev);
788 for (i = 0; i < vmd->msix_count; i++)
789 devm_free_irq(dev, pci_irq_vector(pdev, i), &vmd->irqs[i]);
797 struct vmd_dev *vmd = pci_get_drvdata(pdev);
800 for (i = 0; i < vmd->msix_count; i++) {
803 "vmd", &vmd->irqs[i]);
833 .name = "vmd",