Lines Matching refs:mmu_lock
76 cond_resched_rwlock_write(&kvm->mmu_lock);
105 if (need_resched() || rwlock_needbreak(&kvm->mmu_lock))
122 lockdep_assert_held_write(&kvm->mmu_lock);
134 write_unlock(&kvm->mmu_lock);
140 write_lock(&kvm->mmu_lock);
315 * be called while holding mmu_lock (unless for freeing the stage2 pgd before
325 lockdep_assert_held_write(&kvm->mmu_lock);
359 write_lock(&kvm->mmu_lock);
365 write_unlock(&kvm->mmu_lock);
998 write_lock(&kvm->mmu_lock);
1004 write_unlock(&kvm->mmu_lock);
1014 write_lock(&kvm->mmu_lock);
1021 write_unlock(&kvm->mmu_lock);
1087 write_lock(&kvm->mmu_lock);
1090 write_unlock(&kvm->mmu_lock);
1137 write_lock(&kvm->mmu_lock);
1139 write_unlock(&kvm->mmu_lock);
1149 * Acquires kvm->mmu_lock. Called with kvm->slots_lock mutex acquired,
1166 write_lock(&kvm->mmu_lock);
1168 write_unlock(&kvm->mmu_lock);
1180 * splits them to PAGE_SIZE. Caller must acquire kvm->mmu_lock.
1190 lockdep_assert_held_write(&kvm->mmu_lock);
1310 * was successful and we are holding the mmu_lock, so if this
1499 * acquiring kvm->mmu_lock.
1539 read_lock(&kvm->mmu_lock);
1605 read_unlock(&kvm->mmu_lock);
1618 read_lock(&vcpu->kvm->mmu_lock);
1621 read_unlock(&vcpu->kvm->mmu_lock);
2087 write_lock(&kvm->mmu_lock);
2089 write_unlock(&kvm->mmu_lock);