Lines Matching defs:nr_pages

589 	int nr_pages = 1 << order;
592 for (i = 1; i < nr_pages; i++)
1213 int nr_pages;
1223 nr_pages = 1 << order;
1232 count -= nr_pages;
1233 pcp->count -= nr_pages;
1296 unsigned int nr_pages = 1 << order;
1306 for (loop = 0; loop < (nr_pages - 1); loop++, p++) {
1314 atomic_long_add(nr_pages, &page_zone(page)->managed_pages);
4262 * @nr_pages: The number of pages desired on the list or array
4267 * allocate nr_pages quickly. Pages are added to page_list if page_list
4270 * For lists, nr_pages is the number of pages that should be allocated.
4272 * For arrays, only NULL elements are populated with pages and nr_pages
4278 nodemask_t *nodemask, int nr_pages,
4297 while (page_array && nr_populated < nr_pages && page_array[nr_populated])
4301 if (unlikely(nr_pages <= 0))
4305 if (unlikely(page_array && nr_pages - nr_populated == 0))
4313 if (nr_pages - nr_populated == 1)
4349 mark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK) + nr_pages;
4372 while (nr_populated < nr_pages) {
6275 unsigned long nr_pages, gfp_t gfp_mask)
6277 unsigned long end_pfn = start_pfn + nr_pages;
6284 unsigned long nr_pages)
6286 unsigned long i, end_pfn = start_pfn + nr_pages;
6307 unsigned long start_pfn, unsigned long nr_pages)
6309 unsigned long last_pfn = start_pfn + nr_pages - 1;
6316 * @nr_pages: Number of contiguous pages to allocate
6326 * The allocated memory is always aligned to a page boundary. If nr_pages is a
6328 * nr_pages (e.g. 1GB request would be aligned to 1GB).
6335 struct page *alloc_contig_pages(unsigned long nr_pages, gfp_t gfp_mask,
6348 pfn = ALIGN(zone->zone_start_pfn, nr_pages);
6349 while (zone_spans_last_pfn(zone, pfn, nr_pages)) {
6350 if (pfn_range_valid_contig(zone, pfn, nr_pages)) {
6359 ret = __alloc_contig_pages(pfn, nr_pages,
6365 pfn += nr_pages;
6373 void free_contig_range(unsigned long pfn, unsigned long nr_pages)
6377 for (; nr_pages--; pfn++) {