Lines Matching refs:size

64 				"#size-cells", NULL);
95 static void *unflatten_dt_alloc(void **mem, unsigned long size,
102 *mem += size;
285 * Return: The size of unflattened device tree or error code
369 void *(*dt_alloc)(u64 size, u64 align),
372 int size;
384 pr_debug("size: %08x\n", fdt_totalsize(blob));
392 /* First pass, scan for size */
393 size = unflatten_dt_nodes(blob, NULL, dad, NULL);
394 if (size < 0)
397 size = ALIGN(size, 4);
398 pr_debug(" size is %d, allocating...\n", size);
401 mem = dt_alloc(size + 4, __alignof__(struct device_node));
405 memset(mem, 0, size);
407 *(__be32 *)(mem + size) = cpu_to_be32(0xdeadbeef);
413 if (be32_to_cpup(mem + size) != 0xdeadbeef)
415 be32_to_cpup(mem + size));
426 static void *kernel_tree_alloc(u64 size, u64 align)
428 return kzalloc(size, GFP_KERNEL);
479 phys_addr_t base, size;
499 size = dt_mem_next_cell(dt_root_size_cells, &prop);
501 if (size &&
502 early_init_dt_reserve_memory_arch(base, size, nomap) == 0)
503 pr_debug("Reserved memory: reserved region for node '%s': base %pa, size %lu MiB\n",
504 uname, &base, (unsigned long)(size / SZ_1M));
506 pr_info("Reserved memory: failed to reserve memory for node '%s': base %pa, size %lu MiB\n",
507 uname, &base, (unsigned long)(size / SZ_1M));
511 fdt_reserved_mem_save_node(node, uname, base, size);
519 * __reserved_mem_check_root() - check if #size-cells, #address-cells provided
527 prop = of_get_flat_dt_prop(node, "#size-cells", NULL);
571 if (err == -ENOENT && of_get_flat_dt_prop(node, "size", NULL))
588 u64 base, size;
595 fdt_get_mem_rsv(initial_boot_params, n, &base, &size);
596 if (!size)
598 early_init_dt_reserve_memory_arch(base, size, false);
706 int *size)
708 return fdt_getprop(initial_boot_params, node, name, size);
829 int size;
833 prop = of_get_flat_dt_prop(dt_root, "compatible", &size);
835 while (size > 0) {
837 size -= strlen(prop) + 1;
953 * early_init_dt_scan_root - fetch the top level address and size cells
966 prop = of_get_flat_dt_prop(node, "#size-cells", NULL);
1012 pr_debug("memory scan node %s, reg size %d,\n", uname, l);
1015 u64 base, size;
1018 size = dt_mem_next_cell(dt_root_size_cells, &reg);
1020 if (size == 0)
1023 (unsigned long long)size);
1025 early_init_dt_add_memory_arch(base, size);
1030 if (early_init_dt_mark_hotplug_memory_arch(base, size))
1032 base, base + size);
1101 void __init __weak early_init_dt_add_memory_arch(u64 base, u64 size)
1105 if (size < PAGE_SIZE - (base & ~PAGE_MASK)) {
1107 base, base + size);
1112 size -= PAGE_SIZE - (base & ~PAGE_MASK);
1115 size &= PAGE_MASK;
1119 base, base + size);
1123 if (base + size - 1 > MAX_MEMBLOCK_ADDR) {
1125 ((u64)MAX_MEMBLOCK_ADDR) + 1, base + size);
1126 size = MAX_MEMBLOCK_ADDR - base + 1;
1129 if (base + size < phys_offset) {
1131 base, base + size);
1137 size -= phys_offset - base;
1140 memblock_add(base, size);
1143 int __init __weak early_init_dt_mark_hotplug_memory_arch(u64 base, u64 size)
1145 return memblock_mark_hotplug(base, size);
1149 phys_addr_t size, bool nomap)
1156 if (memblock_is_region_reserved(base, size))
1159 return memblock_mark_nomap(base, size);
1161 return memblock_reserve(base, size);
1164 static void * __init early_init_dt_alloc_memory_arch(u64 size, u64 align)
1166 void *ptr = memblock_alloc(size, align);
1170 __func__, size, align);
1201 /* Initialize {size,address}-cells info */
1252 int size;
1260 size = fdt_totalsize(initial_boot_params);
1261 dt = early_init_dt_alloc_memory_arch(size,
1265 memcpy(dt, initial_boot_params, size);
1293 of_fdt_raw_attr.size = fdt_totalsize(initial_boot_params);