Lines Matching refs:size
14 * for free if we use the same page size as the 1-1 mappings. In that
41 unsigned long size,
45 return memblock_alloc_try_nid_raw(size, align, goal,
49 void * __meminit vmemmap_alloc_block(unsigned long size, int node)
54 int order = get_order(size);
69 return __earlyonly_bootmem_alloc(node, size, size,
73 static void * __meminit altmap_alloc_block_buf(unsigned long size,
76 /* need to make sure size is all the same during early stage */
77 void * __meminit vmemmap_alloc_block_buf(unsigned long size, int node,
83 return altmap_alloc_block_buf(size, altmap);
85 ptr = sparse_buffer_alloc(size);
87 ptr = vmemmap_alloc_block(size, node);
106 static void * __meminit altmap_alloc_block_buf(unsigned long size,
111 if (size & ~PAGE_MASK) {
113 __func__, size);
118 nr_pfns = size >> PAGE_SHIFT;
176 static void * __meminit vmemmap_alloc_block_zero(unsigned long size, int node)
178 void *p = vmemmap_alloc_block(size, node);
182 memset(p, 0, size);
349 * a configuration issue with the size of the altmap.
363 * For compound pages bigger than section size (e.g. x86 1G compound
364 * pages with 2M subsection size) fill the rest of sections as tail
404 unsigned long size, addr;
421 size = min(end - start, pgmap_vmemmap_nr(pgmap) * sizeof(struct page));
422 for (addr = start; addr < end; addr += size) {
423 unsigned long next, last = addr + size;