Lines Matching refs:addr
57 #define __TLBI_VADDR(addr, asid) \
59 unsigned long __ta = (addr) >> 12; \
104 #define __tlbi_level(op, addr, level) do { \
105 u64 arg = addr; \
136 #define __TLBI_VADDR_RANGE(addr, asid, scale, num, ttl) \
138 unsigned long __ta = (addr) >> PAGE_SHIFT; \
200 * flush_tlb_page(vma, addr)
201 * Invalidate a single user mapping for address 'addr' in the
213 * __flush_tlb_kernel_pgtable(addr)
214 * Invalidate a single kernel mapping for address 'addr' on all
260 unsigned long addr;
263 addr = __TLBI_VADDR(uaddr, ASID(vma->vm_mm));
264 __tlbi(vale1is, addr);
265 __tlbi_user(vale1is, addr);
288 unsigned long asid, addr, pages;
321 * ((num+1)*2^(5*scale+1) starting from 'addr'), then increase it
331 addr = __TLBI_VADDR(start, asid);
333 __tlbi_level(vale1is, addr, tlb_level);
334 __tlbi_user_level(vale1is, addr, tlb_level);
336 __tlbi_level(vae1is, addr, tlb_level);
337 __tlbi_user_level(vae1is, addr, tlb_level);
346 addr = __TLBI_VADDR_RANGE(start, asid, scale,
349 __tlbi(rvale1is, addr);
350 __tlbi_user(rvale1is, addr);
352 __tlbi(rvae1is, addr);
353 __tlbi_user(rvae1is, addr);
376 unsigned long addr;
387 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12))
388 __tlbi(vaale1is, addr);
399 unsigned long addr = __TLBI_VADDR(kaddr, 0);
402 __tlbi(vaae1is, addr);