Lines Matching defs:start
461 * We start only with one section of pages, more pages are added as
1497 void __meminit reserve_bootmem_region(phys_addr_t start, phys_addr_t end)
1499 unsigned long start_pfn = PFN_DOWN(start);
1921 unsigned long start = jiffies;
1946 * pre-grown prior to start of deferred page initialization.
1969 .start = spfn,
1985 pgdat->node_id, jiffies_to_msecs(jiffies - start));
4886 /* Reclaim has failed us, start killing things */
5218 /* reset page count bias and offset to start of new frag */
5243 /* reset page count bias and offset to start of new frag */
5453 * start swapping. Assume at least half of the page cache, or the
6163 * made on memory-hotadd so a system can start with mobility
6280 unsigned long start = jiffies;
6289 * the end of that region and start processing the device pages.
6335 nr_pages, jiffies_to_msecs(jiffies - start));
6534 /* start with a fail safe value for batch */
6680 * get_pfn_range_for_nid - Return the start and end page frames for a node
6685 * It returns the start and end page frame of a node based on information
6687 * with no available memory, a warning is printed and the start and end
6783 /* Get the start and end of the zone */
6824 * @start_pfn: The start PFN to start searching for holes
7178 unsigned long __maybe_unused start = 0;
7185 start = pgdat->node_start_pfn & ~(MAX_ORDER_NR_PAGES - 1);
7186 offset = pgdat->node_start_pfn - start;
7199 size = (end - start) * sizeof(struct page);
7300 unsigned long start, end, mask;
7304 for_each_mem_pfn_range(i, MAX_NUMNODES, &start, &end, &nid) {
7305 if (!start || last_nid < 0 || last_nid == nid) {
7313 * start pfn and tick off bits one-by-one until it becomes
7316 mask = ~((1 << __ffs(start)) - 1);
7317 while (mask && last_end <= (start & (mask << 1)))
7563 /* Align start of ZONE_MOVABLE on all nids to MAX_ORDER_NR_PAGES */
7672 pr_info("Movable zone start for each node\n");
7775 unsigned long free_reserved_area(void *start, void *end, int poison, const char *s)
7780 start = (void *)PAGE_ALIGN((unsigned long)start);
7782 for (pos = start; pos < end; pos += PAGE_SIZE, pages++) {
7838 #define adj_init_size(start, end, size, pos, adj) \
7840 if (&start[0] <= &pos[0] && &pos[0] < &end[0] && size > adj) \
8588 /* [start, end) must belong to a single zone. */
8590 unsigned long start, unsigned long end)
8594 unsigned long pfn = start;
8639 * @start: start PFN to allocate
8655 * pages which PFN is in [start, end) are allocated for the caller and
8658 int alloc_contig_range(unsigned long start, unsigned long end,
8668 .zone = page_zone(pfn_to_page(start)),
8701 ret = start_isolate_page_range(pfn_max_align_down(start),
8716 ret = __alloc_contig_migrate_range(&cc, start, end);
8722 * Pages from [start, end) are within a MAX_ORDER_NR_PAGES
8724 * more, all pages in [start, end) are free in page allocator.
8726 * [start, end) (that is remove them from page allocator).
8741 outer_start = start;
8744 outer_start = start;
8750 if (outer_start != start) {
8755 * it doesn't include start page. Adjust outer_start
8759 if (outer_start + (1UL << order) <= start)
8760 outer_start = start;
8779 if (start != outer_start)
8780 free_contig_range(outer_start, start - outer_start);
8785 undo_isolate_page_range(pfn_max_align_down(start),