/kernel/linux/linux-5.10/arch/microblaze/mm/ |
H A D | pgtable.c | 60 p = addr & PAGE_MASK; in __ioremap() 117 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap() 130 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap() 202 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr() 204 p4d = p4d_offset(pgd, addr & PAGE_MASK); in get_pteptr() 205 pud = pud_offset(p4d, addr & PAGE_MASK); in get_pteptr() 206 pmd = pmd_offset(pud, addr & PAGE_MASK); in get_pteptr() 208 pte = pte_offset_kernel(pmd, addr & PAGE_MASK); in get_pteptr() 238 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); in iopa() [all...] |
/kernel/linux/linux-6.6/arch/microblaze/mm/ |
H A D | pgtable.c | 62 p = addr & PAGE_MASK; in __ioremap() 119 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap() 132 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap() 204 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr() 206 p4d = p4d_offset(pgd, addr & PAGE_MASK); in get_pteptr() 207 pud = pud_offset(p4d, addr & PAGE_MASK); in get_pteptr() 208 pmd = pmd_offset(pud, addr & PAGE_MASK); in get_pteptr() 210 pte = pte_offset_kernel(pmd, addr & PAGE_MASK); in get_pteptr() 240 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); in iopa() [all...] |
/kernel/linux/linux-5.10/arch/m68k/mm/ |
H A D | cache.c | 50 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 68 return (*descaddr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 88 address &= PAGE_MASK; in flush_icache_user_range()
|
/kernel/linux/linux-6.6/arch/arc/include/asm/ |
H A D | pgtable-levels.h | 108 #define p4d_bad(x) ((p4d_val(x) & ~PAGE_MASK)) 111 #define p4d_pgtable(p4d) ((pud_t *)(p4d_val(p4d) & PAGE_MASK)) 130 #define pud_bad(x) ((pud_val(x) & ~PAGE_MASK)) 133 #define pud_pgtable(pud) ((pmd_t *)(pud_val(pud) & PAGE_MASK)) 157 #define pmd_bad(x) ((pmd_val(x) & ~PAGE_MASK)) 160 #define pmd_page_vaddr(pmd) (pmd_val(pmd) & PAGE_MASK) 161 #define pmd_pfn(pmd) ((pmd_val(pmd) & PAGE_MASK) >> PAGE_SHIFT)
|
/kernel/linux/linux-5.10/arch/m68k/kernel/ |
H A D | sys_m68k.c | 63 _paddr = (_mmusr & MMU_R_040) ? (_mmusr & PAGE_MASK) : 0; \ 104 paddr += addr & ~(PAGE_MASK | 15); in cache_flush_040() 107 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_040() 125 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_040() 180 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_040() 268 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_060() 286 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_060() 319 addr &= PAGE_MASK; in cache_flush_060() 341 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_060() 342 addr &= PAGE_MASK; /* Workaroun in cache_flush_060() [all...] |
/kernel/linux/linux-6.6/arch/m68k/kernel/ |
H A D | sys_m68k.c | 63 _paddr = (_mmusr & MMU_R_040) ? (_mmusr & PAGE_MASK) : 0; \ 104 paddr += addr & ~(PAGE_MASK | 15); in cache_flush_040() 107 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_040() 125 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_040() 180 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_040() 268 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_060() 286 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_060() 319 addr &= PAGE_MASK; in cache_flush_060() 341 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_060() 342 addr &= PAGE_MASK; /* Workaroun in cache_flush_060() [all...] |
/kernel/linux/linux-6.6/arch/mips/mm/ |
H A D | tlb-r3k.c | 87 start &= PAGE_MASK; in local_flush_tlb_range() 89 end &= PAGE_MASK; in local_flush_tlb_range() 123 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 125 end &= PAGE_MASK; in local_flush_tlb_kernel_range() 160 page &= PAGE_MASK; in local_flush_tlb_page() 201 address &= PAGE_MASK; in __update_tlb()
|
/kernel/linux/linux-5.10/arch/sh/mm/ |
H A D | tlbflush_32.c | 25 page &= PAGE_MASK; in local_flush_tlb_page() 60 start &= PAGE_MASK; in local_flush_tlb_range() 62 end &= PAGE_MASK; in local_flush_tlb_range() 93 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 95 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
|
/kernel/linux/linux-6.6/arch/sh/mm/ |
H A D | tlbflush_32.c | 25 page &= PAGE_MASK; in local_flush_tlb_page() 60 start &= PAGE_MASK; in local_flush_tlb_range() 62 end &= PAGE_MASK; in local_flush_tlb_range() 93 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 95 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
H A D | doorbell.c | 56 if (page->user_virt == (virt & PAGE_MASK)) in mlx5_ib_db_map_user() 65 page->user_virt = (virt & PAGE_MASK); in mlx5_ib_db_map_user() 67 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx5_ib_db_map_user() 78 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); in mlx5_ib_db_map_user()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
H A D | doorbell.c | 56 if (page->user_virt == (virt & PAGE_MASK)) in mlx4_ib_db_map_user() 65 page->user_virt = (virt & PAGE_MASK); in mlx4_ib_db_map_user() 67 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx4_ib_db_map_user() 78 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); in mlx4_ib_db_map_user()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
H A D | fbmem.h | 63 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_peek() 64 u32 val = ioread32(p + (off & ~PAGE_MASK)); in fbmem_peek() 72 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_poke() 73 iowrite32(val, p + (off & ~PAGE_MASK)); in fbmem_poke()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/ |
H A D | doorbell.c | 58 (page->user_virt == (virt & PAGE_MASK))) in mlx5_ib_db_map_user() 67 page->user_virt = (virt & PAGE_MASK); in mlx5_ib_db_map_user() 69 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx5_ib_db_map_user() 83 (virt & ~PAGE_MASK); in mlx5_ib_db_map_user()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx4/ |
H A D | doorbell.c | 56 if (page->user_virt == (virt & PAGE_MASK)) in mlx4_ib_db_map_user() 65 page->user_virt = (virt & PAGE_MASK); in mlx4_ib_db_map_user() 67 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx4_ib_db_map_user() 79 (virt & ~PAGE_MASK); in mlx4_ib_db_map_user()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
H A D | fbmem.h | 63 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_peek() 64 u32 val = ioread32(p + (off & ~PAGE_MASK)); in fbmem_peek() 72 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_poke() 73 iowrite32(val, p + (off & ~PAGE_MASK)); in fbmem_poke()
|
/kernel/linux/linux-5.10/arch/powerpc/mm/ |
H A D | ioremap_64.c | 24 paligned = addr & PAGE_MASK; in __ioremap_caller() 25 offset = addr & ~PAGE_MASK; in __ioremap_caller() 57 addr = (void *)((unsigned long __force)PCI_FIX_ADDR(token) & PAGE_MASK); in iounmap()
|
/kernel/linux/linux-6.6/arch/mips/lib/ |
H A D | r3k_dump_tlb.c | 44 if ((entryhi & PAGE_MASK) != KSEG0 && in dump_tlb() 54 entryhi & PAGE_MASK, in dump_tlb() 56 entrylo0 & PAGE_MASK, in dump_tlb()
|
/kernel/linux/linux-5.10/arch/microblaze/pci/ |
H A D | indirect_pci.c | 148 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci() 152 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci() 153 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci() 154 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci() 155 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
|
/kernel/linux/linux-5.10/arch/powerpc/sysdev/ |
H A D | indirect_pci.c | 163 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci() 167 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci() 168 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci() 169 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci() 170 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
|
/kernel/linux/linux-6.6/arch/powerpc/sysdev/ |
H A D | indirect_pci.c | 162 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci() 166 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci() 167 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci() 168 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci() 169 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
|
/kernel/linux/linux-6.6/arch/powerpc/kernel/ |
H A D | isa-bridge.c | 44 WARN_ON_ONCE(ISA_IO_BASE & ~PAGE_MASK); in remap_isa_base() 45 WARN_ON_ONCE(pa & ~PAGE_MASK); in remap_isa_base() 46 WARN_ON_ONCE(size & ~PAGE_MASK); in remap_isa_base() 78 if ((range.bus_addr & ~PAGE_MASK) || (range.cpu_addr & ~PAGE_MASK)) { in process_ISA_OF_ranges()
|
/kernel/linux/linux-5.10/arch/mips/mm/ |
H A D | tlb-r3k.c | 89 start &= PAGE_MASK; in local_flush_tlb_range() 91 end &= PAGE_MASK; in local_flush_tlb_range() 125 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 127 end &= PAGE_MASK; in local_flush_tlb_kernel_range() 162 page &= PAGE_MASK; in local_flush_tlb_page() 203 address &= PAGE_MASK; in __update_tlb()
|
/kernel/linux/linux-5.10/arch/xtensa/mm/ |
H A D | ioremap.c | 17 unsigned long offset = paddr & ~PAGE_MASK; in xtensa_ioremap() 23 paddr &= PAGE_MASK; in xtensa_ioremap() 61 void *addr = (void *)(PAGE_MASK & (unsigned long)io_addr); in xtensa_iounmap()
|
/kernel/linux/linux-5.10/arch/mips/lib/ |
H A D | r3k_dump_tlb.c | 48 if ((entryhi & PAGE_MASK) != KSEG0 && in dump_tlb() 58 entryhi & PAGE_MASK, in dump_tlb() 60 entrylo0 & PAGE_MASK, in dump_tlb()
|
/kernel/linux/linux-6.6/arch/m68k/mm/ |
H A D | cache.c | 50 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 71 address &= PAGE_MASK; in flush_icache_user_range()
|