Lines Matching defs:page

3  * mm/page-writeback.c
249 * available for the page cache, which is essentially the number of
270 * page cache. This is the base value for the per-node dirty limits.
343 * page cache. This is the base value for the global dirty limits.
845 * when sleeping max_pause per page is not enough to keep the dirty pages under
1943 * In theory 1 page is enough to keep the consumer-producer
1944 * pipe going: the flusher cleans 1 page => the task dirties 1
1945 * more page. However wb_dirty has accounting errors. So use
1967 * called to throttle the page dirties. The solution is to save the not yet
1968 * throttled page dirties in dirty_throttle_leaks on task exit and charge them
1980 * Processes which are dirtying memory should call in here once for each page
2054 * Processes which are dirtying memory should call in here once for each page
2299 * Called early on to tune the page writeback dirty limits.
2330 * @start: starting page index
2331 * @end: ending page index (inclusive)
2333 * This function scans the page range from @start to @end (inclusive) and tags
2346 void *page;
2349 xas_for_each_marked(&xas, page, end, PAGECACHE_TAG_DIRTY) {
2367 * @writepage: function called for each page
2370 * If a page is already under I/O, write_cache_pages() skips it, even
2383 * by the process clearing the DIRTY tag (and submitting the page for IO).
2387 * we do not loop back to the start of the file. Doing so causes a page
2388 * lock/page writeback access order inversion - we should only ever lock
2389 * multiple pages in ascending page->index order, and looping back to the start
2445 * then, even for data integrity operations: the page
2448 * even if there is now a new, dirty page at the same
2481 * past this page so media errors won't choke
2485 * still have state to clear for each page. In
2520 * If we hit the last page and there is more work to be done: wrap
2537 int ret = mapping->a_ops->writepage(&folio->page, wbc);
2634 * Helper function for deaccounting dirty page without writeback.
2649 * Mark the folio dirty, and set it dirty in the page cache, and mark
2657 * means (eg zap_vma_pages() has it mapped and is holding the page table
2757 * the page table lock for a page table which contains at least one page
2758 * in this folio. Truncation will block on the page table lock as it
2790 * page->mapping->host, and if the page is unlocked. This is because another
2791 * CPU could truncate the page off the mapping and then free the mapping.
2793 * Usually, the page _is_ locked, or the caller is a user-space process which
2796 * In other cases, the page should be locked before running set_page_dirty().
2798 int set_page_dirty_lock(struct page *page)
2802 lock_page(page);
2803 ret = set_page_dirty(page);
2804 unlock_page(page);
2810 * This cancels just the dirty bit on the kernel page itself, it does NOT
2812 * leaves the page tagged dirty, so any sync activity will still find it on
2816 * Doing this should *normally* only ever be done when a page is truncated,
2819 * page without actually doing it through the VM. Can you say "ext3 is
2902 * page lock while dirtying the folio, and folios are
3043 * If writeback has been triggered on a page that cannot be made