Lines Matching refs:addr

459  *	if (get_user_pages(addr, 1, FOLL_WRITE, &page, NULL) == 1)
463 * in case the application has unmapped and remapped mm,addr meanwhile.
471 static int break_ksm(struct vm_area_struct *vma, unsigned long addr)
478 page = follow_page(vma, addr,
483 ret = handle_mm_fault(vma, addr,
522 unsigned long addr)
527 vma = find_vma(mm, addr);
528 if (!vma || vma->vm_start > addr)
538 unsigned long addr = rmap_item->address;
548 vma = find_mergeable_vma(mm, addr);
550 break_ksm(vma, addr);
557 unsigned long addr = rmap_item->address;
562 vma = find_mergeable_vma(mm, addr);
566 page = follow_page(vma, addr, FOLL_GET);
570 flush_anon_page(vma, page, addr);
849 unsigned long addr;
852 for (addr = start; addr < end && !err; addr += PAGE_SIZE) {
858 err = break_ksm(vma, addr);
1029 void *addr = kmap_atomic(page);
1030 checksum = xxhash(addr, PAGE_SIZE, 0);
1031 kunmap_atomic(addr);
1130 unsigned long addr;
1134 addr = page_address_in_vma(page, vma);
1135 if (addr == -EFAULT)
1138 pmd = mm_find_pmd(mm, addr);
1142 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma, mm, addr,
1143 addr + PAGE_SIZE);
1146 ptep = pte_offset_map_lock(mm, pmd, addr, &ptl);
1158 page_add_anon_rmap(kpage, vma, addr, false);
1172 flush_cache_page(vma, addr, pte_pfn(*ptep));
1179 ptep_clear_flush(vma, addr, ptep);
1180 set_pte_at_notify(mm, addr, ptep, newpte);
2204 unsigned long addr)
2210 if ((rmap_item->address & PAGE_MASK) == addr)
2212 if (rmap_item->address > addr)
2223 rmap_item->address = addr;
2636 unsigned long addr;
2642 addr = rmap_item->address & ~KSM_FLAG_MASK;
2644 if (addr < vma->vm_start || addr >= vma->vm_end)
2658 if (!rwc->rmap_one(page, vma, addr, rwc->arg)) {