Lines Matching refs:addr

135 	unsigned long addr;
137 for (addr = start; addr <= end; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
138 const pgd_t *pgd_ref = pgd_offset_k(addr);
142 if (addr < start)
153 pgd = (pgd_t *)page_address(page) + pgd_index(addr);
172 unsigned long addr;
174 for (addr = start; addr <= end; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
175 pgd_t *pgd_ref = pgd_offset_k(addr);
184 p4d_ref = p4d_offset(pgd_ref, addr);
195 pgd = (pgd_t *)page_address(page) + pgd_index(addr);
196 p4d = p4d_offset(pgd, addr);
490 pr_info(" pte=%p addr=%lx pte=%016lx\n", pte, paddr,
853 static bool __meminit vmemmap_pmd_is_unused(unsigned long addr, unsigned long end)
855 unsigned long start = ALIGN_DOWN(addr, PMD_SIZE);
862 memset((void *)addr, PAGE_UNUSED, end - addr);
1060 remove_pte_table(pte_t *pte_start, unsigned long addr, unsigned long end,
1067 pte = pte_start + pte_index(addr);
1068 for (; addr < end; addr = next, pte++) {
1069 next = (addr + PAGE_SIZE) & PAGE_MASK;
1081 phys_addr = pte_val(*pte) + (addr & PAGE_MASK);
1089 pte_clear(&init_mm, addr, pte);
1103 remove_pmd_table(pmd_t *pmd_start, unsigned long addr, unsigned long end,
1110 pmd = pmd_start + pmd_index(addr);
1111 for (; addr < end; addr = next, pmd++) {
1112 next = pmd_addr_end(addr, end);
1118 if (IS_ALIGNED(addr, PMD_SIZE) &&
1130 else if (vmemmap_pmd_is_unused(addr, next)) {
1142 remove_pte_table(pte_base, addr, next, direct);
1152 remove_pud_table(pud_t *pud_start, unsigned long addr, unsigned long end,
1159 pud = pud_start + pud_index(addr);
1160 for (; addr < end; addr = next, pud++) {
1161 next = pud_addr_end(addr, end);
1167 IS_ALIGNED(addr, PUD_SIZE) &&
1177 remove_pmd_table(pmd_base, addr, next, direct, altmap);
1186 remove_p4d_table(p4d_t *p4d_start, unsigned long addr, unsigned long end,
1193 p4d = p4d_start + p4d_index(addr);
1194 for (; addr < end; addr = next, p4d++) {
1195 next = p4d_addr_end(addr, end);
1203 remove_pud_table(pud_base, addr, next, altmap, direct);
1223 unsigned long addr;
1227 for (addr = start; addr < end; addr = next) {
1228 next = pgd_addr_end(addr, end);
1230 pgd = pgd_offset_k(addr);
1235 remove_p4d_table(p4d, addr, next, altmap, direct);
1288 unsigned long addr;
1291 for (addr = VMALLOC_START; addr <= VMEMORY_END; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
1292 pgd_t *pgd = pgd_offset_k(addr);
1297 p4d = p4d_alloc(&init_mm, pgd, addr);
1315 pud = pud_alloc(&init_mm, p4d, addr);
1496 unsigned long addr, unsigned long next)
1509 addr_start = addr;
1514 addr_end = addr + PMD_SIZE;
1517 if (!IS_ALIGNED(addr, PMD_SIZE) ||
1519 vmemmap_use_new_sub_pmd(addr, next);
1523 unsigned long addr, unsigned long next)
1528 vmemmap_verify((pte_t *)pmd, node, addr, next);
1529 vmemmap_use_sub_pmd(addr, next);
1562 unsigned long addr = (unsigned long)start_page;
1572 for (; addr < end; addr = next) {
1575 pgd = pgd_offset_k(addr);
1577 next = (addr + PAGE_SIZE) & PAGE_MASK;
1582 p4d = p4d_offset(pgd, addr);
1584 next = (addr + PAGE_SIZE) & PAGE_MASK;
1589 pud = pud_offset(p4d, addr);
1591 next = (addr + PAGE_SIZE) & PAGE_MASK;
1597 next = (addr + PAGE_SIZE) & PAGE_MASK;
1598 pmd = pmd_offset(pud, addr);
1604 pte = pte_offset_kernel(pmd, addr);
1610 next = pmd_addr_end(addr, end);
1612 pmd = pmd_offset(pud, addr);