Lines Matching defs:page

95                struct address_space *mapping, u64 start, struct page **pages,
143 const u8 *data_in, struct page *dest_page,
214 btrfs_page_clamp_clear_writeback(fs_info, &folio->page,
278 struct page **compressed_pages,
312 * full page then submit bio for each compressed/regular extents.
314 * This means, if we have several sectors in the same page points to the same
316 * this function can only help for the next page.
329 struct page *page;
343 * For current subpage support, we only support 64K page size,
344 * which means maximum compressed extent size (128K) is just 2x page
362 page = xa_load(&mapping->i_pages, pg_index);
363 if (page && !xa_is_value(page)) {
372 * Jump to next page start as we already have page for
379 page = __page_cache_alloc(mapping_gfp_constraint(mapping,
381 if (!page)
384 if (add_to_page_cache_lru(page, mapping, pg_index, GFP_NOFS)) {
385 put_page(page);
386 /* There is already a page, skip to page end */
391 if (!*memstall && PageWorkingset(page)) {
396 ret = set_page_extent_mapped(page);
398 unlock_page(page);
399 put_page(page);
410 * At this point, we have a locked page in the page cache for
419 unlock_page(page);
420 put_page(page);
425 if (page->index == end_index) {
431 memzero_page(page, zero_offset, zeros);
436 ret = bio_add_page(orig_bio, page, add_size, offset_in_page(cur));
439 unlock_page(page);
440 put_page(page);
446 * subpage::readers and to unlock the page.
449 btrfs_subpage_start_reader(fs_info, page, cur, add_size);
450 put_page(page);
510 cb->compressed_pages = kcalloc(cb->nr_pages, sizeof(struct page *), GFP_NOFS);
905 u64 start, struct page **pages,
940 * single page, and we want to read a single page out of it.
943 int btrfs_decompress(int type, const u8 *data_in, struct page *dest_page,
1304 struct page *page;
1330 page = find_get_page(inode->i_mapping, index);
1331 in_data = kmap_local_page(page);
1335 /* Don't sample any garbage from the last page */
1345 put_page(page);