Lines Matching refs:dist
42 struct vgic_dist *dist = &kvm->arch.vgic;
65 raw_spin_lock_irqsave(&dist->lpi_list_lock, flags);
71 list_for_each_entry(oldirq, &dist->lpi_list_head, lpi_list) {
89 list_add_tail(&irq->lpi_list, &dist->lpi_list_head);
90 dist->lpi_list_count++;
93 raw_spin_unlock_irqrestore(&dist->lpi_list_lock, flags);
321 struct vgic_dist *dist = &kvm->arch.vgic;
334 irq_count = READ_ONCE(dist->lpi_list_count);
339 raw_spin_lock_irqsave(&dist->lpi_list_lock, flags);
340 list_for_each_entry(irq, &dist->lpi_list_head, lpi_list) {
348 raw_spin_unlock_irqrestore(&dist->lpi_list_lock, flags);
548 static struct vgic_irq *__vgic_its_check_cache(struct vgic_dist *dist,
554 list_for_each_entry(cte, &dist->lpi_translation_cache, entry) {
570 if (!list_is_first(&cte->entry, &dist->lpi_translation_cache))
571 list_move(&cte->entry, &dist->lpi_translation_cache);
582 struct vgic_dist *dist = &kvm->arch.vgic;
586 raw_spin_lock_irqsave(&dist->lpi_list_lock, flags);
588 irq = __vgic_its_check_cache(dist, db, devid, eventid);
592 raw_spin_unlock_irqrestore(&dist->lpi_list_lock, flags);
601 struct vgic_dist *dist = &kvm->arch.vgic;
610 raw_spin_lock_irqsave(&dist->lpi_list_lock, flags);
612 if (unlikely(list_empty(&dist->lpi_translation_cache)))
621 if (__vgic_its_check_cache(dist, db, devid, eventid))
625 cte = list_last_entry(&dist->lpi_translation_cache,
644 list_move(&cte->entry, &dist->lpi_translation_cache);
647 raw_spin_unlock_irqrestore(&dist->lpi_list_lock, flags);
652 struct vgic_dist *dist = &kvm->arch.vgic;
656 raw_spin_lock_irqsave(&dist->lpi_list_lock, flags);
658 list_for_each_entry(cte, &dist->lpi_translation_cache, entry) {
670 raw_spin_unlock_irqrestore(&dist->lpi_list_lock, flags);
1842 struct vgic_dist *dist = &kvm->arch.vgic;
1846 if (!list_empty(&dist->lpi_translation_cache))
1860 list_add(&cte->entry, &dist->lpi_translation_cache);
1866 struct vgic_dist *dist = &kvm->arch.vgic;
1872 &dist->lpi_translation_cache, entry) {