Lines Matching refs:vaddr
158 static void __init __sme_early_map_unmap_mem(void *vaddr, unsigned long size,
161 unsigned long paddr = (unsigned long)vaddr - __PAGE_OFFSET;
169 __early_make_pgtable((unsigned long)vaddr, pmd);
171 vaddr += PMD_SIZE;
292 static void enc_dec_hypercall(unsigned long vaddr, unsigned long size, bool enc)
295 unsigned long vaddr_end = vaddr + size;
297 while (vaddr < vaddr_end) {
302 kpte = lookup_address(vaddr, &level);
304 WARN_ONCE(1, "kpte lookup for vaddr\n");
317 vaddr = (vaddr & pmask) + psize;
322 static bool amd_enc_status_change_prepare(unsigned long vaddr, int npages, bool enc)
329 snp_set_memory_shared(vaddr, npages);
335 static bool amd_enc_status_change_finish(unsigned long vaddr, int npages, bool enc)
342 snp_set_memory_private(vaddr, npages);
345 enc_dec_hypercall(vaddr, npages << PAGE_SHIFT, enc);
405 static int __init early_set_memory_enc_dec(unsigned long vaddr,
414 start = vaddr;
415 vaddr_next = vaddr;
416 vaddr_end = vaddr + size;
418 for (; vaddr < vaddr_end; vaddr = vaddr_next) {
419 kpte = lookup_address(vaddr, &level);
427 vaddr_next = (vaddr & PAGE_MASK) + PAGE_SIZE;
440 if (vaddr == (vaddr & pmask) &&
441 ((vaddr_end - vaddr) >= psize)) {
443 vaddr_next = (vaddr & pmask) + psize;
462 kernel_physical_mapping_change(__pa(vaddr & pmask),
475 int __init early_set_memory_decrypted(unsigned long vaddr, unsigned long size)
477 return early_set_memory_enc_dec(vaddr, size, false);
480 int __init early_set_memory_encrypted(unsigned long vaddr, unsigned long size)
482 return early_set_memory_enc_dec(vaddr, size, true);
485 void __init early_set_mem_enc_dec_hypercall(unsigned long vaddr, unsigned long size, bool enc)
487 enc_dec_hypercall(vaddr, size, enc);
535 unsigned long vaddr, vaddr_end, npages;
538 vaddr = (unsigned long)__start_bss_decrypted_unused;
540 npages = (vaddr_end - vaddr) >> PAGE_SHIFT;
551 r = set_memory_encrypted(vaddr, npages);
558 free_init_pages("unused decrypted", vaddr, vaddr_end);