Lines Matching refs:size

36 						    unsigned long size,
100 __early_ioremap(resource_size_t phys_addr, unsigned long size, pgprot_t prot)
119 __func__, &phys_addr, size))
122 /* Don't allow wraparound or zero size */
123 last_addr = phys_addr + size - 1;
124 if (WARN_ON(!size || last_addr < phys_addr))
127 prev_size[slot] = size;
133 size = PAGE_ALIGN(last_addr + 1) - phys_addr;
138 nrpages = size >> PAGE_SHIFT;
156 __func__, &phys_addr, size, slot, offset, slot_virt[slot]);
162 void __init early_iounmap(void __iomem *addr, unsigned long size)
179 __func__, addr, size))
182 if (WARN(prev_size[slot] != size,
183 "%s(%p, %08lx) [%d] size not consistent %08lx\n",
184 __func__, addr, size, slot, prev_size[slot]))
188 __func__, addr, size, slot);
195 nrpages = PAGE_ALIGN(offset + size) >> PAGE_SHIFT;
211 early_ioremap(resource_size_t phys_addr, unsigned long size)
213 return __early_ioremap(phys_addr, size, FIXMAP_PAGE_IO);
218 early_memremap(resource_size_t phys_addr, unsigned long size)
220 pgprot_t prot = early_memremap_pgprot_adjust(phys_addr, size,
223 return (__force void *)__early_ioremap(phys_addr, size, prot);
227 early_memremap_ro(resource_size_t phys_addr, unsigned long size)
229 pgprot_t prot = early_memremap_pgprot_adjust(phys_addr, size,
232 return (__force void *)__early_ioremap(phys_addr, size, prot);
238 early_memremap_prot(resource_size_t phys_addr, unsigned long size,
241 return (__force void *)__early_ioremap(phys_addr, size,
248 void __init copy_from_early_mem(void *dest, phys_addr_t src, unsigned long size)
253 while (size) {
255 clen = size;
263 size -= clen;
270 early_ioremap(resource_size_t phys_addr, unsigned long size)
277 early_memremap(resource_size_t phys_addr, unsigned long size)
282 early_memremap_ro(resource_size_t phys_addr, unsigned long size)
287 void __init early_iounmap(void __iomem *addr, unsigned long size)
294 void __init early_memunmap(void *addr, unsigned long size)
296 early_iounmap((__force void __iomem *)addr, size);