Lines Matching defs:end
356 * and smp_mb in walk_shadow_page_lockless_begin/end.
547 unsigned long end);
553 unsigned long end;
592 if (WARN_ON_ONCE(range->end <= range->start))
607 range->start, range->end - 1) {
612 hva_end = min(range->end, slot->userspace_addr +
629 gfn_range.end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, slot);
636 range->on_lock(kvm, range->start, range->end);
661 unsigned long end,
668 .end = end,
682 unsigned long end,
688 .end = end,
703 * surrounded by invalidate_range_{start,end}(), which is currently
727 * .change_pte() must be surrounded by .invalidate_range_{start,end}().
742 unsigned long end)
752 kvm->mmu_invalidate_range_end = end;
766 max(kvm->mmu_invalidate_range_end, end);
776 .end = range->end,
784 trace_kvm_unmap_hva_range(range->start, range->end);
808 gfn_to_pfn_cache_invalidate_start(kvm, range->start, range->end,
817 unsigned long end)
840 .end = range->end,
869 unsigned long end)
871 trace_kvm_age_hva(start, end);
873 return kvm_handle_hva_range(mn, start, end, KVM_MMU_NOTIFIER_NO_ARG,
880 unsigned long end)
882 trace_kvm_age_hva(start, end);
897 return kvm_handle_hva_range_no_flush(mn, start, end, kvm_age_gfn);
1929 gfn_t start, gfn_t end)
1933 kvm_for_each_memslot_in_gfn_range(&iter, slots, start, end) {
3512 ktime_t end, bool success)
3515 u64 poll_ns = ktime_to_ns(ktime_sub(end, start));
3710 * Set at the beginning and cleared at the end of interception/PLE handler.