Lines Matching refs:VADDR
31 * | - | CONTEXT | - | VADDR bits 63:22 |
139 * termination. VADDR will not be clobbered, but REG2 will.
156 #define KERN_PGTABLE_WALK(VADDR, REG1, REG2, FAIL_LABEL) \
159 sllx VADDR, 64 - (PGDIR_SHIFT + PGDIR_BITS), REG2; \
164 sllx VADDR, 64 - (PUD_SHIFT + PUD_BITS), REG2; \
176 sllx VADDR, 64 - (PMD_SHIFT + PMD_BITS), REG2; \
188 and VADDR, REG2, REG2; \
191 698: sllx VADDR, 64 - PMD_SHIFT, REG2; \
208 #define USER_PGTABLE_CHECK_PUD_HUGE(VADDR, REG1, REG2, FAIL_LABEL, PTE_LABEL) \
224 and VADDR, REG2, REG2; \
229 #define USER_PGTABLE_CHECK_PUD_HUGE(VADDR, REG1, REG2, FAIL_LABEL, PTE_LABEL) \
243 #define USER_PGTABLE_CHECK_PMD_HUGE(VADDR, REG1, REG2, FAIL_LABEL, PTE_LABEL) \
252 and VADDR, REG2, REG2; \
257 #define USER_PGTABLE_CHECK_PMD_HUGE(VADDR, REG1, REG2, FAIL_LABEL, PTE_LABEL) \
269 * VADDR will not be clobbered, but REG1 and REG2 will.
271 #define USER_PGTABLE_WALK_TL1(VADDR, PHYS_PGD, REG1, REG2, FAIL_LABEL) \
272 sllx VADDR, 64 - (PGDIR_SHIFT + PGDIR_BITS), REG2; \
277 sllx VADDR, 64 - (PUD_SHIFT + PUD_BITS), REG2; \
281 USER_PGTABLE_CHECK_PUD_HUGE(VADDR, REG1, REG2, FAIL_LABEL, 800f) \
283 sllx VADDR, 64 - (PMD_SHIFT + PMD_BITS), REG2; \
287 USER_PGTABLE_CHECK_PMD_HUGE(VADDR, REG1, REG2, FAIL_LABEL, 800f) \
288 sllx VADDR, 64 - PMD_SHIFT, REG2; \
297 /* Lookup a OBP mapping on VADDR in the prom_trans[] table at TL>0.
299 * the resulting PTE value will be left in REG1. VADDR is preserved
302 #define OBP_TRANS_LOOKUP(VADDR, REG1, REG2, REG3, FAIL_LABEL) \
310 cmp REG2, VADDR; \
312 cmp VADDR, REG3; \
315 sub VADDR, REG2, REG2; \
331 /* Do a kernel TSB lookup at tl>0 on VADDR+TAG, branch to OK_LABEL
336 * VADDR and TAG will be preserved and not clobbered by this macro.
338 #define KERN_TSB_LOOKUP_TL1(VADDR, TAG, REG1, REG2, REG3, REG4, OK_LABEL) \
348 srlx VADDR, PAGE_SHIFT, REG2; \
358 /* This version uses a trick, the TAG is already (VADDR >> 22) so