Lines Matching refs:page_start

47  * @page_start: page index of the first page to be freed
50 * Free pages [@page_start and @page_end) in @pages for all units.
54 struct page **pages, int page_start, int page_end)
60 for (i = page_start; i < page_end; i++) {
73 * @page_start: page index of the first page to be allocated
77 * Allocate pages [@page_start,@page_end) into @pages for all units.
82 struct page **pages, int page_start, int page_end,
91 for (i = page_start; i < page_end; i++) {
102 while (--i >= page_start)
108 for (i = page_start; i < page_end; i++)
117 * @page_start: page index of the first page to be flushed
120 * Pages in [@page_start,@page_end) of @chunk are about to be
127 int page_start, int page_end)
130 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start),
143 * @page_start: page index of the first page to unmap
146 * For each cpu, unmap pages [@page_start,@page_end) out of @chunk.
153 struct page **pages, int page_start, int page_end)
159 for (i = page_start; i < page_end; i++) {
166 __pcpu_unmap_pages(pcpu_chunk_addr(chunk, cpu, page_start),
167 page_end - page_start);
174 * @page_start: page index of the first page to be flushed
177 * Pages [@page_start,@page_end) of @chunk have been unmapped. Flush
185 int page_start, int page_end)
188 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start),
203 * @page_start: page index of the first page to map
206 * For each cpu, map pages [@page_start,@page_end) into @chunk. The
214 struct page **pages, int page_start, int page_end)
220 err = __pcpu_map_pages(pcpu_chunk_addr(chunk, cpu, page_start),
221 &pages[pcpu_page_idx(cpu, page_start)],
222 page_end - page_start);
226 for (i = page_start; i < page_end; i++)
235 __pcpu_unmap_pages(pcpu_chunk_addr(chunk, tcpu, page_start),
236 page_end - page_start);
238 pcpu_post_unmap_tlb_flush(chunk, page_start, page_end);
245 * @page_start: page index of the first page to be flushed
248 * Pages [@page_start,@page_end) of @chunk have been mapped. Flush
255 int page_start, int page_end)
258 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start),
265 * @page_start: the start page
269 * For each cpu, populate and map pages [@page_start,@page_end) into
276 int page_start, int page_end, gfp_t gfp)
284 if (pcpu_alloc_pages(chunk, pages, page_start, page_end, gfp))
287 if (pcpu_map_pages(chunk, pages, page_start, page_end)) {
288 pcpu_free_pages(chunk, pages, page_start, page_end);
291 pcpu_post_map_flush(chunk, page_start, page_end);
299 * @page_start: the start page
302 * For each cpu, depopulate and unmap pages [@page_start,@page_end)
309 int page_start, int page_end)
322 pcpu_pre_unmap_flush(chunk, page_start, page_end);
324 pcpu_unmap_pages(chunk, pages, page_start, page_end);
328 pcpu_free_pages(chunk, pages, page_start, page_end);