Lines Matching refs:eoi
531 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu);
534 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
536 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
541 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu);
552 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
554 elem = list_first_entry_or_null(&eoi->eoi_list, struct irq_info,
557 list_add(&info->eoi_list, &eoi->eoi_list);
559 &eoi->delayed, delay);
561 list_for_each_entry_reverse(elem, &eoi->eoi_list, eoi_list) {
568 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
612 struct lateeoi_work *eoi;
617 eoi = container_of(to_delayed_work(work), struct lateeoi_work, delayed);
622 spin_lock_irqsave(&eoi->eoi_list_lock, flags);
624 info = list_first_entry_or_null(&eoi->eoi_list, struct irq_info,
632 &eoi->delayed,
639 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
646 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags);
653 struct lateeoi_work *eoi = &per_cpu(lateeoi, cpu);
655 INIT_DELAYED_WORK(&eoi->delayed, xen_irq_lateeoi_worker);
656 spin_lock_init(&eoi->eoi_list_lock);
657 INIT_LIST_HEAD(&eoi->eoi_list);
792 struct physdev_eoi eoi = { .irq = pirq_from_irq(data->irq) };
811 rc = HYPERVISOR_physdev_op(PHYSDEVOP_eoi, &eoi);
1013 * interrupts too. In any case Xen always honors the eoi mechanism,
1015 * hasn't received an eoi yet. Therefore using the fasteoi handler