Lines Matching defs:end
1449 unsigned long end,
1471 hva_end = min(end, memslot->userspace_addr +
1505 int kvm_unmap_hva_range(struct kvm *kvm, unsigned long start, unsigned long end,
1510 r = kvm_handle_hva_range(kvm, start, end, 0, kvm_unmap_rmapp);
1513 r |= kvm_tdp_mmu_zap_hva_range(kvm, start, end);
1574 int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned long end)
1578 young = kvm_handle_hva_range(kvm, start, end, 0, kvm_age_rmapp);
1580 young |= kvm_tdp_mmu_age_hva_range(kvm, start, end);
1600 u64 *end;
1602 for (pos = spt, end = pos + PAGE_SIZE / sizeof(u64); pos != end; pos++)
2066 /* The page is good, but __kvm_sync_page might still end
2375 * with vcpu_enter_guest and walk_shadow_page_lockless_begin/end.
2676 u64 *start, u64 *end)
2689 ret = gfn_to_page_many_atomic(slot, gfn, pages, end - start);
5514 gfn_t start, end;
5517 end = min(gfn_end, memslot->base_gfn + memslot->npages);
5518 if (start >= end)
5524 start, end - 1, true);