Lines Matching defs:range
86 * Get the maximum number of page-tables pages needed to split a range
87 * of blocks into PAGE_SIZE PTEs. It assumes the range is already
90 static int kvm_mmu_split_nr_page_tables(u64 range)
95 n += DIV_ROUND_UP(range, PUD_SIZE);
96 n += DIV_ROUND_UP(range, PMD_SIZE);
308 * unmap_stage2_range -- Clear stage2 page table entries to unmap a range
310 * @start: The intermediate physical base address of the range to unmap
314 * Clear a range of stage-2 mappings, lowering the various ref-counts. Must
565 * create_hyp_mappings - duplicate a kernel virtual address range in Hyp mode
566 * @from: The virtual kernel start address of the range
567 * @to: The virtual kernel end address of the range (exclusive)
568 * @prot: The protection to be applied to this range
613 * overflowed the idmap/IO address range.
624 * hyp_alloc_private_va_range - Allocates a private VA range.
625 * @size: The size of the VA range to reserve.
628 * The private virtual address (VA) range is allocated below io_map_base
773 * create_hyp_exec_mappings - Map an executable range into HYP
1056 * kvm_phys_addr_ioremap - map a device range to guest IPA
1102 * stage2_wp_range() - write protect stage2 memory region range
1104 * @addr: Start address of range
1105 * @end: End address of range
1660 /* Falls between the IPA range and the PARange? */
1727 * we know for sure is that this range cannot be cached.
1771 bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range)
1776 __unmap_stage2_range(&kvm->arch.mmu, range->start << PAGE_SHIFT,
1777 (range->end - range->start) << PAGE_SHIFT,
1778 range->may_block);
1783 bool kvm_set_spte_gfn(struct kvm *kvm, struct kvm_gfn_range *range)
1785 kvm_pfn_t pfn = pte_pfn(range->arg.pte);
1790 WARN_ON(range->end - range->start != 1);
1810 kvm_pgtable_stage2_map(kvm->arch.mmu.pgt, range->start << PAGE_SHIFT,
1817 bool kvm_age_gfn(struct kvm *kvm, struct kvm_gfn_range *range)
1819 u64 size = (range->end - range->start) << PAGE_SHIFT;
1825 range->start << PAGE_SHIFT,
1829 bool kvm_test_age_gfn(struct kvm *kvm, struct kvm_gfn_range *range)
1831 u64 size = (range->end - range->start) << PAGE_SHIFT;
1837 range->start << PAGE_SHIFT,
1896 * range. This is only the case if system RAM is out of range for the
1898 * also need the extended virtual range for the HYP ID map, or we won't
1908 * line: we need to use the extended range with *both* our translation
1921 kvm_debug("HYP VA range: %lx:%lx\n",