/kernel/linux/linux-5.10/arch/x86/mm/ |
H A D | mem_encrypt_identity.c | 79 unsigned long vaddr_end; member 105 pgd_end = ppd->vaddr_end & PGDIR_MASK; in sme_clear_pgd() 195 while (ppd->vaddr < ppd->vaddr_end) { in __sme_map_range_pmd() 205 while (ppd->vaddr < ppd->vaddr_end) { in __sme_map_range_pte() 216 unsigned long vaddr_end; in __sme_map_range() local 222 vaddr_end = ppd->vaddr_end; in __sme_map_range() 225 ppd->vaddr_end = ALIGN(ppd->vaddr, PMD_PAGE_SIZE); in __sme_map_range() 229 ppd->vaddr_end = vaddr_end in __sme_map_range() [all...] |
H A D | mem_encrypt.c | 259 unsigned long vaddr_end, vaddr_next; in early_set_memory_enc_dec() local 266 vaddr_end = vaddr + size; in early_set_memory_enc_dec() 268 for (; vaddr < vaddr_end; vaddr = vaddr_next) { in early_set_memory_enc_dec() 291 ((vaddr_end - vaddr) >= psize)) { in early_set_memory_enc_dec() 313 __pa((vaddr_end & pmask) + psize), in early_set_memory_enc_dec() 391 unsigned long vaddr, vaddr_end, npages; in mem_encrypt_free_decrypted_mem() local 395 vaddr_end = (unsigned long)__end_bss_decrypted; in mem_encrypt_free_decrypted_mem() 396 npages = (vaddr_end - vaddr) >> PAGE_SHIFT; in mem_encrypt_free_decrypted_mem() 410 free_init_pages("unused decrypted", vaddr, vaddr_end); in mem_encrypt_free_decrypted_mem()
|
H A D | kaslr.c | 41 static const unsigned long vaddr_end = CPU_ENTRY_AREA_BASE; variable 78 * with the vaddr_start/vaddr_end variables. These checks are very in kernel_randomize_memory() 81 BUILD_BUG_ON(vaddr_start >= vaddr_end); in kernel_randomize_memory() 82 BUILD_BUG_ON(vaddr_end != CPU_ENTRY_AREA_BASE); in kernel_randomize_memory() 83 BUILD_BUG_ON(vaddr_end > __START_KERNEL_map); in kernel_randomize_memory() 112 remain_entropy = vaddr_end - vaddr_start; in kernel_randomize_memory()
|
H A D | init_64.c | 423 unsigned long vaddr_end = __START_KERNEL_map + KERNEL_IMAGE_SIZE; in cleanup_highmap() local 433 vaddr_end = __START_KERNEL_map + (max_pfn_mapped << PAGE_SHIFT); in cleanup_highmap() 435 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap() 677 unsigned long vaddr, vaddr_end, vaddr_next, paddr_next, paddr_last; in phys_p4d_init() local 681 vaddr_end = (unsigned long)__va(paddr_end); in phys_p4d_init() 687 for (; vaddr < vaddr_end; vaddr = vaddr_next) { in phys_p4d_init() 707 paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end), in phys_p4d_init() 713 paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end), in phys_p4d_init() 731 unsigned long vaddr, vaddr_start, vaddr_end, vaddr_next, paddr_last; in __kernel_physical_mapping_init() local 735 vaddr_end in __kernel_physical_mapping_init() [all...] |
/kernel/linux/linux-6.6/arch/x86/mm/ |
H A D | mem_encrypt_identity.c | 82 unsigned long vaddr_end; member 108 pgd_end = ppd->vaddr_end & PGDIR_MASK; in sme_clear_pgd() 198 while (ppd->vaddr < ppd->vaddr_end) { in __sme_map_range_pmd() 208 while (ppd->vaddr < ppd->vaddr_end) { in __sme_map_range_pte() 219 unsigned long vaddr_end; in __sme_map_range() local 225 vaddr_end = ppd->vaddr_end; in __sme_map_range() 228 ppd->vaddr_end = ALIGN(ppd->vaddr, PMD_SIZE); in __sme_map_range() 232 ppd->vaddr_end = vaddr_end in __sme_map_range() [all...] |
H A D | mem_encrypt_amd.c | 295 unsigned long vaddr_end = vaddr + size; in enc_dec_hypercall() local 297 while (vaddr < vaddr_end) { in enc_dec_hypercall() 408 unsigned long vaddr_end, vaddr_next, start; in early_set_memory_enc_dec() local 416 vaddr_end = vaddr + size; in early_set_memory_enc_dec() 418 for (; vaddr < vaddr_end; vaddr = vaddr_next) { in early_set_memory_enc_dec() 441 ((vaddr_end - vaddr) >= psize)) { in early_set_memory_enc_dec() 463 __pa((vaddr_end & pmask) + psize), in early_set_memory_enc_dec() 535 unsigned long vaddr, vaddr_end, npages; in mem_encrypt_free_decrypted_mem() local 539 vaddr_end = (unsigned long)__end_bss_decrypted; in mem_encrypt_free_decrypted_mem() 540 npages = (vaddr_end in mem_encrypt_free_decrypted_mem() [all...] |
H A D | kaslr.c | 41 static const unsigned long vaddr_end = CPU_ENTRY_AREA_BASE; variable 78 * with the vaddr_start/vaddr_end variables. These checks are very in kernel_randomize_memory() 81 BUILD_BUG_ON(vaddr_start >= vaddr_end); in kernel_randomize_memory() 82 BUILD_BUG_ON(vaddr_end != CPU_ENTRY_AREA_BASE); in kernel_randomize_memory() 83 BUILD_BUG_ON(vaddr_end > __START_KERNEL_map); in kernel_randomize_memory() 112 remain_entropy = vaddr_end - vaddr_start; in kernel_randomize_memory()
|
H A D | init_64.c | 429 unsigned long vaddr_end = __START_KERNEL_map + KERNEL_IMAGE_SIZE; in cleanup_highmap() local 439 vaddr_end = __START_KERNEL_map + (max_pfn_mapped << PAGE_SHIFT); in cleanup_highmap() 441 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap() 678 unsigned long vaddr, vaddr_end, vaddr_next, paddr_next, paddr_last; in phys_p4d_init() local 682 vaddr_end = (unsigned long)__va(paddr_end); in phys_p4d_init() 688 for (; vaddr < vaddr_end; vaddr = vaddr_next) { in phys_p4d_init() 708 paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end), in phys_p4d_init() 714 paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end), in phys_p4d_init() 732 unsigned long vaddr, vaddr_start, vaddr_end, vaddr_next, paddr_last; in __kernel_physical_mapping_init() local 736 vaddr_end in __kernel_physical_mapping_init() [all...] |
/kernel/linux/linux-5.10/arch/x86/kernel/ |
H A D | head64.c | 139 unsigned long vaddr, vaddr_end; in __startup_64() local 291 vaddr_end = (unsigned long)__end_bss_decrypted; in __startup_64() 292 for (; vaddr < vaddr_end; vaddr += PMD_SIZE) { in __startup_64()
|
/kernel/linux/linux-6.6/arch/x86/kernel/ |
H A D | head64.c | 132 unsigned long vaddr, vaddr_end; in sme_postprocess_startup() local 146 vaddr_end = (unsigned long)__end_bss_decrypted; in sme_postprocess_startup() 148 for (; vaddr < vaddr_end; vaddr += PMD_SIZE) { in sme_postprocess_startup()
|
H A D | sev.c | 793 unsigned long vaddr_end, int op) in __set_pages_state() 810 while (vaddr < vaddr_end && i < ARRAY_SIZE(data->entries)) { in __set_pages_state() 825 (vaddr_end - vaddr) >= PMD_SIZE) { in __set_pages_state() 867 unsigned long vaddr_end; in set_pages_state() local 874 vaddr_end = vaddr + (npages << PAGE_SHIFT); in set_pages_state() 876 while (vaddr < vaddr_end) in set_pages_state() 877 vaddr = __set_pages_state(&desc, vaddr, vaddr_end, op); in set_pages_state() 792 __set_pages_state(struct snp_psc_desc *data, unsigned long vaddr, unsigned long vaddr_end, int op) __set_pages_state() argument
|
H A D | sev-shared.c | 1092 unsigned long vaddr_end = vaddr + PMD_SIZE; in pvalidate_pages() local 1094 for (; vaddr < vaddr_end; vaddr += PAGE_SIZE) { in pvalidate_pages()
|
/kernel/linux/linux-5.10/arch/x86/kvm/svm/ |
H A D | sev.c | 436 unsigned long vaddr, vaddr_end, next_vaddr, npages, pages, size, i; in sev_launch_update_data() local 455 vaddr_end = vaddr + size; in sev_launch_update_data() 470 for (i = 0; vaddr < vaddr_end; vaddr = next_vaddr, i += pages) { in sev_launch_update_data() 783 unsigned long vaddr, vaddr_end, next_vaddr; in sev_dbg_crypt() local 804 vaddr_end = vaddr + size; in sev_dbg_crypt() 807 for (; vaddr < vaddr_end; vaddr = next_vaddr) { in sev_dbg_crypt()
|
/kernel/linux/linux-6.6/mm/ |
H A D | huge_memory.c | 2973 unsigned long vaddr_end) in split_huge_pages_pid() 2982 vaddr_end &= PAGE_MASK; in split_huge_pages_pid() 3005 pid, vaddr_start, vaddr_end); in split_huge_pages_pid() 3012 for (addr = vaddr_start; addr < vaddr_end; addr += PAGE_SIZE) { in split_huge_pages_pid() 3128 unsigned long vaddr_start, vaddr_end; in split_huge_pages_write() local 3169 ret = sscanf(input_buf, "%d,0x%lx,0x%lx", &pid, &vaddr_start, &vaddr_end); in split_huge_pages_write() 3179 ret = split_huge_pages_pid(pid, vaddr_start, vaddr_end); in split_huge_pages_write() 2972 split_huge_pages_pid(int pid, unsigned long vaddr_start, unsigned long vaddr_end) split_huge_pages_pid() argument
|
/kernel/linux/linux-6.6/arch/x86/kvm/svm/ |
H A D | sev.c | 506 unsigned long vaddr, vaddr_end, next_vaddr, npages, pages, size, i; in sev_launch_update_data() local 521 vaddr_end = vaddr + size; in sev_launch_update_data() 537 for (i = 0; vaddr < vaddr_end; vaddr = next_vaddr, i += pages) { in sev_launch_update_data() 933 unsigned long vaddr, vaddr_end, next_vaddr; in sev_dbg_crypt() local 954 vaddr_end = vaddr + size; in sev_dbg_crypt() 957 for (; vaddr < vaddr_end; vaddr = next_vaddr) { in sev_dbg_crypt()
|
/kernel/linux/linux-5.10/arch/x86/xen/ |
H A D | mmu_pv.c | 998 unsigned long vaddr_end) in xen_cleanhighmap() 1005 for (; vaddr <= vaddr_end && (pmd < (level2_kernel_pgt + PTRS_PER_PMD)); in xen_cleanhighmap() 1023 void *vaddr_end = vaddr + size; in xen_free_ro_pages() local 1025 for (; vaddr < vaddr_end; vaddr += PAGE_SIZE) in xen_free_ro_pages() 997 xen_cleanhighmap(unsigned long vaddr, unsigned long vaddr_end) xen_cleanhighmap() argument
|
/kernel/linux/linux-6.6/arch/x86/xen/ |
H A D | mmu_pv.c | 1013 unsigned long vaddr_end) in xen_cleanhighmap() 1020 for (; vaddr <= vaddr_end && (pmd < (level2_kernel_pgt + PTRS_PER_PMD)); in xen_cleanhighmap() 1038 void *vaddr_end = vaddr + size; in xen_free_ro_pages() local 1040 for (; vaddr < vaddr_end; vaddr += PAGE_SIZE) in xen_free_ro_pages() 1012 xen_cleanhighmap(unsigned long vaddr, unsigned long vaddr_end) xen_cleanhighmap() argument
|