Lines Matching refs:base
36 * A simple test that adds a memory block of a specified base address
48 .base = SZ_1G,
55 memblock_add(r.base, r.size);
57 ASSERT_EQ(rgn->base, r.base);
69 * A simple test that adds a memory block of a specified base address, size,
81 .base = SZ_1M,
88 memblock_add_node(r.base, r.size, 1, MEMBLOCK_HOTPLUG);
90 ASSERT_EQ(rgn->base, r.base);
125 .base = SZ_1G,
129 .base = SZ_1G + SZ_16K,
136 memblock_add(r1.base, r1.size);
137 memblock_add(r2.base, r2.size);
139 ASSERT_EQ(rgn1->base, r1.base);
142 ASSERT_EQ(rgn2->base, r2.base);
155 * with the beginning of r1 (that is r1.base < r2.base + r2.size):
162 * | r1.base
164 * r2.base
166 * Expect to merge the two entries into one region that starts at r2.base
178 .base = SZ_512M,
182 .base = SZ_256M,
188 total_size = (r1.base - r2.base) + r1.size;
191 memblock_add(r1.base, r1.size);
192 memblock_add(r2.base, r2.size);
194 ASSERT_EQ(rgn->base, r2.base);
207 * with the end of r1 (that is r2.base < r1.base + r1.size):
214 * | r2.base
216 * r1.base
218 * Expect to merge the two entries into one region that starts at r1.base
230 .base = SZ_128M,
234 .base = SZ_256M,
240 total_size = (r2.base - r1.base) + r2.size;
243 memblock_add(r1.base, r1.size);
244 memblock_add(r2.base, r2.size);
246 ASSERT_EQ(rgn->base, r1.base);
259 * within the range of r1 (that is r1.base < r2.base &&
260 * r2.base + r2.size < r1.base + r1.size):
267 * r1.base
279 .base = SZ_8M,
283 .base = SZ_16M,
290 memblock_add(r1.base, r1.size);
291 memblock_add(r2.base, r2.size);
293 ASSERT_EQ(rgn->base, r1.base);
311 .base = SZ_16K,
319 memblock_add(r.base, r.size);
320 memblock_add(r.base, r.size);
338 * Expect to merge the three entries into one region that starts at r1.base
350 .base = SZ_1G,
354 .base = SZ_1G + SZ_16K,
358 .base = SZ_1G + SZ_8K,
367 memblock_add(r1.base, r1.size);
368 memblock_add(r2.base, r2.size);
369 memblock_add(r3.base, r3.size);
371 ASSERT_EQ(rgn->base, r1.base);
393 * Expect to add a memory block of size PHYS_ADDR_MAX - r.base. Expect the
404 .base = PHYS_ADDR_MAX - SZ_1M,
410 total_size = PHYS_ADDR_MAX - r.base;
413 memblock_add(r.base, r.size);
415 ASSERT_EQ(rgn->base, r.base);
437 .base = SZ_16K,
441 phys_addr_t base, size = SZ_64;
455 base = PAGE_ALIGN(dummy_physical_memory_base());
458 memblock_add(base, new_memory_regions_size);
460 /* This is the base of small memory block. */
461 base += new_memory_regions_size + gap_size;
470 memblock_add(base, size);
471 base += size + gap_size;
492 memblock_add(r.base, r.size);
493 ASSERT_EQ(memblock.memory.regions[0].base, r.base);
541 * A simple test that marks a memory block of a specified base address
553 .base = SZ_2G,
560 memblock_reserve(r.base, r.size);
562 ASSERT_EQ(rgn->base, r.base);
589 .base = SZ_256M,
593 .base = SZ_512M,
600 memblock_reserve(r1.base, r1.size);
601 memblock_reserve(r2.base, r2.size);
603 ASSERT_EQ(rgn1->base, r1.base);
606 ASSERT_EQ(rgn2->base, r2.base);
620 * r1.base < r2.base + r2.size):
627 * | r1.base
629 * r2.base
631 * Expect to merge two entries into one region that starts at r2.base and
643 .base = SZ_1G,
647 .base = SZ_128M,
653 total_size = (r1.base - r2.base) + r1.size;
656 memblock_reserve(r1.base, r1.size);
657 memblock_reserve(r2.base, r2.size);
659 ASSERT_EQ(rgn->base, r2.base);
673 * r2.base < r1.base + r1.size):
680 * | r2.base
682 * r1.base
684 * Expect to merge two entries into one region that starts at r1.base and
696 .base = SZ_2K,
700 .base = SZ_128K,
706 total_size = (r2.base - r1.base) + r2.size;
709 memblock_reserve(r1.base, r1.size);
710 memblock_reserve(r2.base, r2.size);
712 ASSERT_EQ(rgn->base, r1.base);
726 * (r1.base < r2.base) && (r2.base + r2.size < r1.base + r1.size)):
733 * | r2.base
735 * r1.base
747 .base = SZ_1M,
751 .base = SZ_2M,
758 memblock_reserve(r1.base, r1.size);
759 memblock_reserve(r2.base, r2.size);
761 ASSERT_EQ(rgn->base, r1.base);
780 .base = SZ_16K,
788 memblock_reserve(r.base, r.size);
789 memblock_reserve(r.base, r.size);
808 * r1.base and has size of r1.size + r2.size + r3.size. The region counter and
819 .base = SZ_1G,
823 .base = SZ_1G + SZ_16K,
827 .base = SZ_1G + SZ_8K,
836 memblock_reserve(r1.base, r1.size);
837 memblock_reserve(r2.base, r2.size);
838 memblock_reserve(r3.base, r3.size);
840 ASSERT_EQ(rgn->base, r1.base);
862 * Expect to reserve a memory block of size PHYS_ADDR_MAX - r.base. Expect the
873 .base = PHYS_ADDR_MAX - SZ_1M,
879 total_size = PHYS_ADDR_MAX - r.base;
882 memblock_reserve(r.base, r.size);
884 ASSERT_EQ(rgn->base, r.base);
906 .base = SZ_16K,
959 memblock_reserve(r.base, r.size);
960 ASSERT_EQ(memblock.reserved.regions[0].base, r.base);
1016 * rgn.base
1029 .base = SZ_2K,
1033 .base = SZ_128K,
1040 memblock_add(r1.base, r1.size);
1041 memblock_add(r2.base, r2.size);
1042 memblock_remove(r1.base, r1.size);
1044 ASSERT_EQ(rgn->base, r2.base);
1067 * rgn.base
1078 .base = SZ_512K,
1082 .base = SZ_64M,
1089 memblock_add(r1.base, r1.size);
1090 memblock_remove(r2.base, r2.size);
1092 ASSERT_EQ(rgn->base, r1.base);
1106 * (that is r1.base < r2.base + r2.size):
1116 * | rgn.base
1117 * r1.base
1130 .base = SZ_32M,
1134 .base = SZ_16M,
1140 r1_end = r1.base + r1.size;
1141 r2_end = r2.base + r2.size;
1145 memblock_add(r1.base, r1.size);
1146 memblock_remove(r2.base, r2.size);
1148 ASSERT_EQ(rgn->base, r1.base + r2.base);
1161 * the already existing region r1 (that is r2.base < r1.base + r1.size):
1171 * r1.base
1184 .base = SZ_2M,
1188 .base = SZ_32M,
1194 total_size = r2.base - r1.base;
1197 memblock_add(r1.base, r1.size);
1198 memblock_remove(r2.base, r2.size);
1200 ASSERT_EQ(rgn->base, r1.base);
1214 * (r1.base < r2.base) && (r2.base + r2.size < r1.base + r1.size)):
1224 * r1.base
1226 * Expect that the region is split into two - one that ends at r2.base and
1227 * another that starts at r2.base + r2.size, with appropriate sizes. The
1239 .base = SZ_1M,
1243 .base = SZ_16M,
1249 r1_size = r2.base - r1.base;
1250 r2_size = (r1.base + r1.size) - (r2.base + r2.size);
1254 memblock_add(r1.base, r1.size);
1255 memblock_remove(r2.base, r2.size);
1257 ASSERT_EQ(rgn1->base, r1.base);
1260 ASSERT_EQ(rgn2->base, r2.base + r2.size);
1285 .base = SZ_2K,
1292 memblock_add(r1.base, r1.size);
1293 memblock_remove(r1.base, r1.size);
1295 ASSERT_EQ(rgn->base, 0);
1317 * Expect that only the portion between PHYS_ADDR_MAX and r2.base is removed.
1329 .base = PHYS_ADDR_MAX - SZ_2M,
1334 .base = PHYS_ADDR_MAX - SZ_1M,
1340 total_size = r1.size - (PHYS_ADDR_MAX - r2.base);
1343 memblock_add(r1.base, r1.size);
1344 memblock_remove(r2.base, r2.size);
1346 ASSERT_EQ(rgn->base, r1.base);
1381 .base = SZ_16M,
1385 .base = SZ_64M,
1389 .base = SZ_32M,
1395 r2_end = r2.base + r2.size;
1396 r3_end = r3.base + r3.size;
1397 new_r1_size = r3.base - r1.base;
1402 memblock_add(r1.base, r1.size);
1403 memblock_add(r2.base, r2.size);
1404 memblock_remove(r3.base, r3.size);
1406 ASSERT_EQ(rgn1->base, r1.base);
1409 ASSERT_EQ(rgn2->base, r3_end);
1450 * rgn.base
1462 .base = SZ_4M,
1466 .base = SZ_8M,
1473 memblock_reserve(r1.base, r1.size);
1474 memblock_reserve(r2.base, r2.size);
1475 memblock_free((void *)r1.base, r1.size);
1477 ASSERT_EQ(rgn->base, r2.base);
1500 * rgn.base
1511 .base = SZ_2M,
1515 .base = SZ_16M,
1522 memblock_reserve(r1.base, r1.size);
1523 memblock_free((void *)r2.base, r2.size);
1525 ASSERT_EQ(rgn->base, r1.base);
1538 * of the already existing entry r1 (that is r1.base < r2.base + r2.size):
1548 * | rgn.base
1550 * r1.base
1563 .base = SZ_8M,
1567 .base = SZ_1M,
1573 total_size = (r1.size + r1.base) - (r2.base + r2.size);
1576 memblock_reserve(r1.base, r1.size);
1577 memblock_free((void *)r2.base, r2.size);
1579 ASSERT_EQ(rgn->base, r2.base + r2.size);
1592 * the already existing entry r1 (that is r2.base < r1.base + r1.size):
1612 .base = SZ_8M,
1616 .base = SZ_32M,
1622 total_size = r2.base - r1.base;
1625 memblock_reserve(r1.base, r1.size);
1626 memblock_free((void *)r2.base, r2.size);
1628 ASSERT_EQ(rgn->base, r1.base);
1642 * (r1.base < r2.base) && (r2.base + r2.size < r1.base + r1.size)):
1652 * r1.base
1654 * Expect that the region is split into two - one that ends at r2.base and
1655 * another that starts at r2.base + r2.size, with appropriate sizes. The
1667 .base = SZ_1M,
1671 .base = SZ_4M,
1677 r1_size = r2.base - r1.base;
1678 r2_size = (r1.base + r1.size) - (r2.base + r2.size);
1682 memblock_reserve(r1.base, r1.size);
1683 memblock_free((void *)r2.base, r2.size);
1685 ASSERT_EQ(rgn1->base, r1.base);
1688 ASSERT_EQ(rgn2->base, r2.base + r2.size);
1713 .base = SZ_2K,
1720 memblock_reserve(r1.base, r1.size);
1721 memblock_free((void *)r1.base, r1.size);
1723 ASSERT_EQ(rgn->base, 0);
1744 * Expect that only the portion between PHYS_ADDR_MAX and r2.base is freed.
1756 .base = PHYS_ADDR_MAX - SZ_2M,
1761 .base = PHYS_ADDR_MAX - SZ_1M,
1767 total_size = r1.size - (PHYS_ADDR_MAX - r2.base);
1770 memblock_reserve(r1.base, r1.size);
1771 memblock_free((void *)r2.base, r2.size);
1773 ASSERT_EQ(rgn->base, r1.base);
1808 .base = SZ_16M,
1812 .base = SZ_64M,
1816 .base = SZ_32M,
1822 r2_end = r2.base + r2.size;
1823 r3_end = r3.base + r3.size;
1824 new_r1_size = r3.base - r1.base;
1829 memblock_reserve(r1.base, r1.size);
1830 memblock_reserve(r2.base, r2.size);
1831 memblock_free((void *)r3.base, r3.size);
1833 ASSERT_EQ(rgn1->base, r1.base);
1836 ASSERT_EQ(rgn2->base, r3_end);
1924 .base = alignment,
1931 memblock_add(r.base, r.size);
1934 ASSERT_EQ(rgn->base, r.base);
1970 .base = alignment,
1974 .base = alignment * 4,
1981 memblock_add(r1.base, r1.size);
1982 memblock_add(r2.base, r2.size);
1985 ASSERT_EQ(rgn->base, r1.base);
1997 * r2. Region r1 is aligned on both ends and region r2 is unaligned at the base
2011 * Expect that r1 will not be trimmed and r2 will be trimmed at the base.
2025 .base = alignment,
2029 .base = alignment * 4 + offset,
2035 new_r2_base = r2.base + (alignment - offset);
2039 memblock_add(r1.base, r1.size);
2040 memblock_add(r2.base, r2.size);
2043 ASSERT_EQ(rgn1->base, r1.base);
2046 ASSERT_EQ(rgn2->base, new_r2_base);
2058 * r2. Region r1 is aligned on both ends and region r2 is aligned at the base
2086 .base = alignment,
2090 .base = alignment * 4,
2099 memblock_add(r1.base, r1.size);
2100 memblock_add(r2.base, r2.size);
2103 ASSERT_EQ(rgn1->base, r1.base);
2106 ASSERT_EQ(rgn2->base, r2.base);