Lines Matching defs:pfn
183 static bool kvm_is_device_pfn(unsigned long pfn)
185 return !pfn_is_map_memory(pfn);
431 u64 pfn;
439 static struct hyp_shared_pfn *find_shared_pfn(u64 pfn, struct rb_node ***node,
449 if (this->pfn < pfn)
451 else if (this->pfn > pfn)
460 static int share_pfn_hyp(u64 pfn)
467 this = find_shared_pfn(pfn, &node, &parent);
479 this->pfn = pfn;
483 ret = kvm_call_hyp_nvhe(__pkvm_host_share_hyp, pfn, 1);
490 static int unshare_pfn_hyp(u64 pfn)
497 this = find_shared_pfn(pfn, &node, &parent);
509 ret = kvm_call_hyp_nvhe(__pkvm_host_unshare_hyp, pfn, 1);
519 u64 pfn;
539 pfn = __phys_to_pfn(cur);
540 ret = share_pfn_hyp(pfn);
551 u64 pfn;
559 pfn = __phys_to_pfn(cur);
560 WARN_ON(unshare_pfn_hyp(pfn));
1285 kvm_pfn_t pfn = *pfnp;
1316 * to PG_head and switch the pfn from a tail page to the head
1320 kvm_release_pfn_clean(pfn);
1321 pfn &= ~(PTRS_PER_PMD - 1);
1322 get_page(pfn_to_page(pfn));
1323 *pfnp = pfn;
1372 static void sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn,
1376 struct page *page = pfn_to_page(pfn);
1408 kvm_pfn_t pfn;
1507 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL,
1509 if (pfn == KVM_PFN_ERR_HWPOISON) {
1513 if (is_error_noslot_pfn(pfn))
1516 if (kvm_is_device_pfn(pfn)) {
1554 hva, &pfn,
1566 sanitise_mte_tags(kvm, pfn, vma_pagesize);
1593 __pfn_to_phys(pfn), prot,
1600 kvm_set_pfn_dirty(pfn);
1606 kvm_release_pfn_clean(pfn);
1785 kvm_pfn_t pfn = pte_pfn(range->arg.pte);
1797 if (kvm_has_mte(kvm) && !page_mte_tagged(pfn_to_page(pfn)))
1811 PAGE_SIZE, __pfn_to_phys(pfn),