Lines Matching defs:direct
699 if (!sp->role.direct)
707 if (!sp->role.direct) {
713 pr_err_ratelimited("gfn mismatch under direct page %llx "
1630 if (!sp->role.direct)
1662 static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, int direct)
1668 if (!direct)
1820 if ((_sp)->gfn != (_gfn) || (_sp)->role.direct) {} else
2024 int direct,
2039 role.direct = direct;
2040 if (role.direct)
2088 sp = kvm_mmu_alloc_page(vcpu, direct);
2093 if (!direct) {
2202 * For the direct sp, if the guest pte's dirty bit
2312 if (!sp->role.invalid && !sp->role.direct)
2708 WARN_ON(!sp->role.direct);
2990 WARN_ON(!sp->role.direct);
3009 * The gfn of direct spte is stable since it is
3109 * Currently, fast page fault only works for direct mapping
3218 u8 level, bool direct)
3228 sp = kvm_mmu_get_page(vcpu, gfn, gva, level, direct, ACC_ALL);
3268 /* root_pgd is ignored for direct MMUs. */
3472 static bool mmio_info_in_cache(struct kvm_vcpu *vcpu, u64 addr, bool direct)
3481 if (direct)
3568 static int handle_mmio_page_fault(struct kvm_vcpu *vcpu, u64 addr, bool direct)
3573 if (mmio_info_in_cache(vcpu, addr, direct))
3587 if (direct)
3834 return (role.direct || pgd == root->pgd) &&
3921 * If this is a direct root page, it doesn't have a write flooding
3924 if (!new_role.direct)
4187 * the direct page table on host, use as much mmu features as
4505 role.base.direct = true;
4578 role.base.direct = !is_paging(vcpu);
4623 role.base.direct = false;
4660 role.base.direct = false;
4732 * shadow pages of their own and so "direct" has no meaning. Set it
4735 role.base.direct = true;
5070 bool direct = vcpu->arch.mmu->direct_map;
5077 r = handle_mmio_page_fault(vcpu, cr2_or_gpa, direct);
5118 if (!mmio_info_in_cache(vcpu, cr2_or_gpa, direct) && !is_guest_mode(vcpu))
5592 if (sp->role.direct && !kvm_is_reserved_pfn(pfn) &&