Lines Matching defs:start

141  * Note, the meta data fields are deliberately placed at the start of the
1296 * @gfn_offset: start of the BITS_PER_LONG pages we care about
1329 * @gfn_offset: start of the BITS_PER_LONG pages we care about
1372 * Huge pages are NOT write protected when we start dirty logging in
1381 gfn_t start = slot->base_gfn + gfn_offset + __ffs(mask);
1385 kvm_mmu_try_split_huge_pages(kvm, slot, start, end, PG_LEVEL_4K);
1387 kvm_mmu_slot_gfn_write_protect(kvm, slot, start, PG_LEVEL_2M);
1390 if (ALIGN(start << PAGE_SHIFT, PMD_SIZE) !=
1569 range->start, range->end - 1, &iterator)
2973 u64 *start, u64 *end)
2981 gfn = kvm_mmu_page_get_gfn(sp, spte_index(start));
2986 ret = gfn_to_page_many_atomic(slot, gfn, pages, end - start);
2990 for (i = 0; i < ret; i++, gfn++, start++) {
2991 mmu_set_spte(vcpu, slot, start, access, gfn,
3002 u64 *spte, *start = NULL;
3012 if (!start)
3014 if (direct_pte_prefetch_many(vcpu, sp, start, spte) < 0)
3016 start = NULL;
3017 } else if (!start)
3018 start = spte;
3020 if (start)
3021 direct_pte_prefetch_many(vcpu, sp, start, spte);
6210 gfn_t start, end;
6221 start = max(gfn_start, memslot->base_gfn);
6223 if (WARN_ON_ONCE(start >= end))
6228 start, end - 1, true, flush);
6501 gfn_t start, gfn_t end,
6514 level, level, start, end - 1, true, false);
6520 u64 start, u64 end,
6527 kvm_shadow_mmu_try_split_huge_pages(kvm, memslot, start, end, target_level);
6529 kvm_tdp_mmu_try_split_huge_pages(kvm, memslot, start, end, target_level, false);
6541 u64 start = memslot->base_gfn;
6542 u64 end = start + memslot->npages;
6549 kvm_shadow_mmu_try_split_huge_pages(kvm, memslot, start, end, target_level);
6554 kvm_tdp_mmu_try_split_huge_pages(kvm, memslot, start, end, target_level, true);