Lines Matching refs:eoi
561 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu);
564 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
566 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
571 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu);
582 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
584 elem = list_first_entry_or_null(&eoi->eoi_list, struct irq_info,
587 list_add(&info->eoi_list, &eoi->eoi_list);
589 &eoi->delayed, delay);
591 list_for_each_entry_reverse(elem, &eoi->eoi_list, eoi_list) {
598 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
654 struct lateeoi_work *eoi;
659 eoi = container_of(to_delayed_work(work), struct lateeoi_work, delayed);
664 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
666 info = list_first_entry_or_null(&eoi->eoi_list, struct irq_info,
674 &eoi->delayed,
681 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
688 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
695 struct lateeoi_work *eoi = &per_cpu(lateeoi, cpu);
697 INIT_DELAYED_WORK(&eoi->delayed, xen_irq_lateeoi_worker);
698 spin_lock_init(&eoi->eoi_list_lock);
699 INIT_LIST_HEAD(&eoi->eoi_list);
825 struct physdev_eoi eoi = { .irq = pirq_from_irq(data->irq) };
834 rc = HYPERVISOR_physdev_op(PHYSDEVOP_eoi, &eoi);
1053 * interrupts too. In any case Xen always honors the eoi mechanism,
1055 * hasn't received an eoi yet. Therefore using the fasteoi handler